fluent-plugin-azure-loganalytics 0.3.1 → 0.4.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +5 -5
- data/ChangeLog.md +4 -0
- data/README.md +117 -4
- data/VERSION +1 -1
- data/docs/install-tdagent-and-the-plugin-on-rhel.md +68 -0
- data/examples/fluent_csv.conf +23 -0
- data/examples/fluent_typecast.conf +25 -0
- data/fluent-plugin-azure-loganalytics.gemspec +1 -1
- data/img/Azure-LogAnalytics-Output-Image-2.png +0 -0
- data/lib/fluent/plugin/out_azure-loganalytics.rb +3 -1
- data/test/plugin/test_azure_loganalytics.rb +12 -14
- metadata +9 -6
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
|
-
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
2
|
+
SHA256:
|
3
|
+
metadata.gz: 22b06d1614e45b6e4bd69d561fd9e1baa607aab324877c55279d68e6f95cdf89
|
4
|
+
data.tar.gz: 810e6437147632a77ec333e484282c320a75482c889d1cfe59d5e4410eb495e0
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: b17088b54b49a51b04d15545b7e15c16cef996ec4ac7c1ff2c1739e383a15f3064a11db2d7bc763833ef325790935c89dc11ff85589b56da9f3f700e3a69ade4
|
7
|
+
data.tar.gz: f80ef8d11245daf1bafa759dbfd26b3b55dbe6649eaa3bf6389c8a5f78a00c2bb76d5499c94051f5b8b5b7b540dd3249d7ec63885df763382faf8b6f85c25422
|
data/ChangeLog.md
CHANGED
@@ -1,3 +1,7 @@
|
|
1
|
+
## 0.4.0
|
2
|
+
* Add endpoint parameter for sovereign cloud - [PR#8](https://github.com/yokawasa/fluent-plugin-azure-loganalytics/pull/8)
|
3
|
+
* Changed dependency for azure-loganalytics-datacollector-api to `>= 0.1.5` - [PR#8](https://github.com/yokawasa/fluent-plugin-azure-loganalytics/pull/8)
|
4
|
+
|
1
5
|
## 0.3.1
|
2
6
|
|
3
7
|
* Add requirements section - [PR#2](https://github.com/yokawasa/fluent-plugin-azure-loganalytics/pull/2)
|
data/README.md
CHANGED
@@ -11,10 +11,20 @@
|
|
11
11
|
| < 0.3.0 | >= v0.12.0 | >= 1.9 |
|
12
12
|
|
13
13
|
## Installation
|
14
|
+
### Installing gems into system Ruby
|
14
15
|
```
|
15
16
|
$ gem install fluent-plugin-azure-loganalytics
|
16
17
|
```
|
17
18
|
|
19
|
+
### Installing gems into td-agent’s Ruby
|
20
|
+
If you installed td-agent and want to add this custom plugins, use td-agent-gem to install as td-agent has own Ruby so you should install gems into td-agent’s Ruby, not system Ruby:
|
21
|
+
|
22
|
+
```
|
23
|
+
$ /usr/sbin/td-agent-gem install fluent-plugin-azure-loganalytics
|
24
|
+
```
|
25
|
+
Please see also [I installed td-agent and want to add custom plugins. How do I do it?](https://docs.fluentd.org/v0.12/articles/faq#i-installed-td-agent-and-want-to-add-custom-plugins.-how-do-i-do-it?)
|
26
|
+
|
27
|
+
|
18
28
|
## Configuration
|
19
29
|
|
20
30
|
### Azure Log Analytics
|
@@ -33,6 +43,7 @@ Once you have the workspace, get Workspace ID and Shared Key (either Primary Key
|
|
33
43
|
customer_id CUSTOMER_ID # Customer ID aka WorkspaceID String
|
34
44
|
shared_key KEY_STRING # The primary or the secondary Connected Sources client authentication key
|
35
45
|
log_type EVENT_TYPE_NAME # The name of the event type. ex) ApacheAccessLog
|
46
|
+
endpoint myendpoint
|
36
47
|
add_time_field true
|
37
48
|
time_field_name mytime
|
38
49
|
time_format %s
|
@@ -45,6 +56,7 @@ Once you have the workspace, get Workspace ID and Shared Key (either Primary Key
|
|
45
56
|
* **customer\_id (required)** - Your Operations Management Suite workspace ID
|
46
57
|
* **shared\_key (required)** - The primary or the secondary Connected Sources client authentication key
|
47
58
|
* **log\_type (required)** - The name of the event type that is being submitted to Log Analytics. log_type only supports alpha characters
|
59
|
+
* **endpoint (optional)** - Default:'ods.opinsights.azure.com'. The service endpoint. You may want to use this param in case of sovereign cloud that has a different endpoint from the public cloud
|
48
60
|
* **time\_generated\_field (optional)** - Default:''(empty string) The name of the time generated field. Be carefule that the value of field should strictly follow the ISO 8601 format (YYYY-MM-DDThh:mm:ssZ). See also [this](https://docs.microsoft.com/en-us/azure/log-analytics/log-analytics-data-collector-api#create-a-request) for more details
|
49
61
|
* **add\_time\_field (optional)** - Default:true. This option allows to insert a time field to record
|
50
62
|
* **time\_field\_name (optional)** - Default:time. This is required only when add_time_field is true
|
@@ -59,7 +71,7 @@ Once you have the workspace, get Workspace ID and Shared Key (either Primary Key
|
|
59
71
|
fluent-plugin-azure-loganalytics adds **time** and **tag** attributes by default if **add_time_field** and **add_tag_field** are true respectively. Below are two types of the plugin configurations - Default and All options configuration.
|
60
72
|
|
61
73
|
### (1) Default Configuration (No options)
|
62
|
-
<u>
|
74
|
+
<u>fluent_1.conf</u>
|
63
75
|
```
|
64
76
|
<source>
|
65
77
|
@type tail # input plugin
|
@@ -78,7 +90,33 @@ fluent-plugin-azure-loganalytics adds **time** and **tag** attributes by default
|
|
78
90
|
```
|
79
91
|
|
80
92
|
### (2) Configuration with All Options
|
81
|
-
<u>
|
93
|
+
<u>fluent_2.conf</u>
|
94
|
+
```
|
95
|
+
<source>
|
96
|
+
@type tail # input plugin
|
97
|
+
path /var/log/apache2/access.log # monitoring file
|
98
|
+
pos_file /tmp/fluentd_pos_file # position file
|
99
|
+
format apache # format
|
100
|
+
tag azure-loganalytics.access # tag
|
101
|
+
</source>
|
102
|
+
|
103
|
+
<match azure-loganalytics.**>
|
104
|
+
@type azure-loganalytics
|
105
|
+
customer_id 818f7bbc-8034-4cc3-b97d-f068dd4cd658
|
106
|
+
shared_key ppC5500KzCcDsOKwM1yWUvZydCuC3m+ds/2xci0byeQr1G3E0Jkygn1N0Rxx/yVBUrDE2ok3vf4ksCzvBmQXHw==(dummy)
|
107
|
+
log_type ApacheAccessLog
|
108
|
+
add_time_field true
|
109
|
+
time_field_name mytime
|
110
|
+
time_format %s
|
111
|
+
localtime true
|
112
|
+
add_tag_field true
|
113
|
+
tag_field_name mytag
|
114
|
+
</match>
|
115
|
+
```
|
116
|
+
### (3) Configuration with Typecast filter
|
117
|
+
|
118
|
+
You want to add typecast filter when you want to cast fields type. The filed type of code and size are cast by typecast filter.
|
119
|
+
<u>fluent_typecast.conf</u>
|
82
120
|
```
|
83
121
|
<source>
|
84
122
|
@type tail # input plugin
|
@@ -88,6 +126,11 @@ fluent-plugin-azure-loganalytics adds **time** and **tag** attributes by default
|
|
88
126
|
tag azure-loganalytics.access # tag
|
89
127
|
</source>
|
90
128
|
|
129
|
+
<filter **>
|
130
|
+
@type typecast
|
131
|
+
types host:string,user:string,method:string,path:string,referer:string,agent:string,code:integer,size:integer
|
132
|
+
</filter>
|
133
|
+
|
91
134
|
<match azure-loganalytics.**>
|
92
135
|
@type azure-loganalytics
|
93
136
|
customer_id 818f7bbc-8034-4cc3-b97d-f068dd4cd658
|
@@ -101,6 +144,54 @@ fluent-plugin-azure-loganalytics adds **time** and **tag** attributes by default
|
|
101
144
|
tag_field_name mytag
|
102
145
|
</match>
|
103
146
|
```
|
147
|
+
[note] you need to install [fluent-plugin-filter-typecast](https://github.com/sonots/fluent-plugin-filter_typecast) for the sample configuration above.
|
148
|
+
```
|
149
|
+
gem install fluent-plugin-filter_typecast
|
150
|
+
```
|
151
|
+
### (4) Configuration with CSV format as input and specific field type as output
|
152
|
+
You want to send to Log Analytics, logs generated with known delimiter (like comma, semi-colon) then you can use the csv format of fluentd and the keys/types properties.
|
153
|
+
This can be used with any log, here implemented with Nginx custom log.
|
154
|
+
<u>fluent_csv.conf</u>
|
155
|
+
|
156
|
+
Suppose your log is formated the way below in the /etc/nginx/conf.d/log.conf:
|
157
|
+
```
|
158
|
+
log_format appcustomlog '"$time_iso8601";"$hostname";$bytes_sent;$request_time;$upstream_response_length;$upstream_response_time;$content_length;"$remote_addr";$status;"$host";"$request";"$http_user_agent"';
|
159
|
+
```
|
160
|
+
And this log is activated throught the /etc/nginx/conf.d/virtualhost.conf :
|
161
|
+
```
|
162
|
+
server {
|
163
|
+
...
|
164
|
+
access_log /var/log/nginx/access.log appcustomlog;
|
165
|
+
...
|
166
|
+
}
|
167
|
+
```
|
168
|
+
You can use the following configuration for the source to tail the log file and format it with proper field type.
|
169
|
+
```
|
170
|
+
<source>
|
171
|
+
@type tail
|
172
|
+
path /var/log/nginx/access.log
|
173
|
+
pos_file /var/log/td-agent/access.log.pos
|
174
|
+
tag nginx.accesslog
|
175
|
+
format csv
|
176
|
+
delimiter ;
|
177
|
+
keys time,hostname,bytes_sent,request_time,content_length,remote_addr,status,host,request,http_user_agent
|
178
|
+
types time:time,hostname:string,bytes_sent:float,request_time:float,content_length:string,remote_addr:string,status:integer,host:string,request:string,http_user_agent:string
|
179
|
+
time_key time
|
180
|
+
time_format %FT%T%z
|
181
|
+
</source>
|
182
|
+
|
183
|
+
<match nginx.accesslog>
|
184
|
+
@type azure-loganalytics
|
185
|
+
customer_id 818f7bbc-8034-4cc3-b97d-f068dd4cd658
|
186
|
+
shared_key ppC5500KzCcDsOKwM1yWUvZydCuC3m+ds/2xci0byeQr1G3E0Jkygn1N0Rxx/yVBUrDE2ok3vf4ksCzvBmQXHw==(dummy)
|
187
|
+
log_type NginxAcessLog
|
188
|
+
time_generated_field time
|
189
|
+
time_format %FT%T%z
|
190
|
+
add_tag_field true
|
191
|
+
tag_field_name mytag
|
192
|
+
</match>
|
193
|
+
```
|
194
|
+
|
104
195
|
|
105
196
|
## Sample inputs and expected records
|
106
197
|
|
@@ -117,9 +208,19 @@ The output record for sample input can be seen at Log Analytics portal like this
|
|
117
208
|
|
118
209
|
![fluent-plugin-azure-loganalytics output image](https://github.com/yokawasa/fluent-plugin-azure-loganalytics/raw/master/img/Azure-LogAnalytics-Output-Image.png)
|
119
210
|
|
211
|
+
<u>Sample Input (nginx custom access log)</u>
|
212
|
+
```
|
213
|
+
"2017-12-13T11:31:59+00:00";"nginx0001";21381;0.238;20882;0.178;-;"193.192.35.178";200;"mynginx.domain.com";"GET /mysite/picture.jpeg HTTP/1.1";"Mozilla/5.0 (Windows NT 10.0; Win64; x64) Chrome/63.0.3239.84 Safari/537.36"
|
214
|
+
```
|
215
|
+
|
216
|
+
<u>Output Record</u>
|
217
|
+
|
218
|
+
Part of the output record for sample input can be seen at Log Analytics portal like this with field of type _s (string) or _d (double):
|
219
|
+
|
220
|
+
![fluent-plugin-azure-loganalytics output image](https://github.com/yokawasa/fluent-plugin-azure-loganalytics/raw/master/img/Azure-LogAnalytics-Output-Image-2.png)
|
120
221
|
|
121
222
|
## Tests
|
122
|
-
### Running test code
|
223
|
+
### Running test code (using System rake)
|
123
224
|
```
|
124
225
|
$ git clone https://github.com/yokawasa/fluent-plugin-azure-loganalytics.git
|
125
226
|
$ cd fluent-plugin-azure-loganalytics
|
@@ -131,6 +232,18 @@ $ vi test/plugin/test_azure_loganalytics.rb
|
|
131
232
|
$ rake test
|
132
233
|
```
|
133
234
|
|
235
|
+
### Running test code (using td-agent's rake)
|
236
|
+
```
|
237
|
+
$ git clone https://github.com/yokawasa/fluent-plugin-azure-loganalytics.git
|
238
|
+
$ cd fluent-plugin-azure-loganalytics
|
239
|
+
|
240
|
+
# edit CONFIG params of test/plugin/test_azure_loganalytics.rb
|
241
|
+
$ vi test/plugin/test_azure_loganalytics.rb
|
242
|
+
|
243
|
+
# run test
|
244
|
+
$ /opt/td-agent/embedded/bin/rake test
|
245
|
+
```
|
246
|
+
|
134
247
|
### Creating package, running and testing locally
|
135
248
|
```
|
136
249
|
$ rake build
|
@@ -148,9 +261,9 @@ $ ab -n 5 -c 2 http://localhost/test/foo.html
|
|
148
261
|
|
149
262
|
## Links
|
150
263
|
|
151
|
-
* http://yokawasa.github.io/fluent-plugin-azure-loganalytics
|
152
264
|
* https://rubygems.org/gems/fluent-plugin-azure-loganalytics
|
153
265
|
* https://rubygems.org/gems/azure-loganalytics-datacollector-api
|
266
|
+
* [How to install td-agent and luent-plugin-azure-loganalytics plugin on RHEL](docs/install-tdagent-and-the-plugin-on-rhel.md)
|
154
267
|
|
155
268
|
## Contributing
|
156
269
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
0.
|
1
|
+
0.4.0
|
@@ -0,0 +1,68 @@
|
|
1
|
+
# How to install td-agent and luent-plugin-azure-loganalytics plugin on RHEL
|
2
|
+
|
3
|
+
This is a quick installation procedure of td-agent and the custom plugin (fluent-plugin-azure-loganalytics) on Red Hat Enterprise Linux (7.4)
|
4
|
+
|
5
|
+
$ cat /etc/os-release
|
6
|
+
```
|
7
|
+
NAME="Red Hat Enterprise Linux Server"
|
8
|
+
VERSION="7.4 (Maipo)"
|
9
|
+
ID="rhel"
|
10
|
+
ID_LIKE="fedora"
|
11
|
+
VARIANT="Server"
|
12
|
+
VARIANT_ID="server"
|
13
|
+
VERSION_ID="7.4"
|
14
|
+
PRETTY_NAME="Red Hat Enterprise Linux Server 7.4 (Maipo)"
|
15
|
+
ANSI_COLOR="0;31"
|
16
|
+
CPE_NAME="cpe:/o:redhat:enterprise_linux:7.4:GA:server"
|
17
|
+
HOME_URL="https://www.redhat.com/"
|
18
|
+
BUG_REPORT_URL="https://bugzilla.redhat.com/"
|
19
|
+
|
20
|
+
REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 7"
|
21
|
+
REDHAT_BUGZILLA_PRODUCT_VERSION=7.4
|
22
|
+
REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux"
|
23
|
+
REDHAT_SUPPORT_PRODUCT_VERSION="7.4"
|
24
|
+
```
|
25
|
+
|
26
|
+
## 0. prerequisites (for Redhat/Centos)
|
27
|
+
Install GCC and Development Tools on a CentOS / RHEL 7 server
|
28
|
+
```
|
29
|
+
$ suod yum group install "Development Tools"
|
30
|
+
```
|
31
|
+
|
32
|
+
## 1. Install td-agent (fluentd)
|
33
|
+
|
34
|
+
Following the [fluentd official page](https://docs.fluentd.org/v0.12/articles/install-by-rpm), install like this:
|
35
|
+
|
36
|
+
```
|
37
|
+
$ curl -L https://toolbelt.treasuredata.com/sh/install-redhat-td-agent2.sh | sh
|
38
|
+
|
39
|
+
$ td-agent --version
|
40
|
+
td-agent 0.12.40
|
41
|
+
```
|
42
|
+
|
43
|
+
## 2. Launching Daemon
|
44
|
+
```
|
45
|
+
$ sudo /etc/init.d/td-agent start
|
46
|
+
$ sudo /etc/init.d/td-agent status
|
47
|
+
```
|
48
|
+
## 3. Post Sample Logs via HTTP
|
49
|
+
By default, /etc/td-agent/td-agent.conf is configured to take logs from HTTP and route them to stdout (/var/log/td-agent/td-agent.log). You can post sample log records using the curl command.
|
50
|
+
|
51
|
+
```
|
52
|
+
$ curl -X POST -d 'json={"json":"message"}' http://localhost:8888/debug.test
|
53
|
+
|
54
|
+
# Checking log (/var/log/td-agent/td-agent.log) and see if the log is written
|
55
|
+
$ cat /var/log/td-agent/td-agent.log
|
56
|
+
```
|
57
|
+
|
58
|
+
## 4. Install the custom plugin
|
59
|
+
```
|
60
|
+
$ sudo /usr/sbin/td-agent-gem install fluent-plugin-azure-loganalytics
|
61
|
+
```
|
62
|
+
|
63
|
+
## 5. Testing the plugin
|
64
|
+
```
|
65
|
+
$ git clone https://github.com/yokawasa/fluent-plugin-azure-loganalytics.git
|
66
|
+
$ cd fluent-plugin-azure-loganalytics
|
67
|
+
$ /opt/td-agent/embedded/bin/rake test
|
68
|
+
```
|
@@ -0,0 +1,23 @@
|
|
1
|
+
<source>
|
2
|
+
@type tail # input plugin
|
3
|
+
path /var/log/nginx/access.log # monitoring file
|
4
|
+
pos_file /var/log/td-agent/access.log.pos # position file
|
5
|
+
format csv # format
|
6
|
+
tag nginx.accesslog # tag
|
7
|
+
delimiter ; # record delimiter used in source log
|
8
|
+
keys time,hostname,bytes_sent,request_time,content_length,remote_addr,status,host,request,http_user_agent
|
9
|
+
types time:time,hostname:string,bytes_sent:float,request_time:float,content_length:string,remote_addr:string,status:integer,host:string,request:string,http_user_agent:string
|
10
|
+
time_key time
|
11
|
+
time_format %FT%T%z
|
12
|
+
</source>
|
13
|
+
|
14
|
+
<match nginx.accesslog>
|
15
|
+
@type azure-loganalytics
|
16
|
+
customer_id CUSTOMER_ID # Customer ID aka WorkspaceID String
|
17
|
+
shared_key KEY_STRING # The primary or the secondary Connected Sources client authentication key
|
18
|
+
log_type EVENT_TYPE_NAME # The name of the event type. ex) NginxAcessLog
|
19
|
+
time_generated_field time
|
20
|
+
time_format %FT%T%z
|
21
|
+
add_tag_field true
|
22
|
+
tag_field_name mytag
|
23
|
+
</match>
|
@@ -0,0 +1,25 @@
|
|
1
|
+
<source>
|
2
|
+
@type tail # input plugin
|
3
|
+
path /var/log/apache2/access.log # monitoring file
|
4
|
+
pos_file /tmp/fluentd_pos_file # position file
|
5
|
+
format apache # format
|
6
|
+
tag azure-loganalytics.access # tag
|
7
|
+
</source>
|
8
|
+
|
9
|
+
<filter **>
|
10
|
+
@type typecast
|
11
|
+
types host:string,user:string,method:string,path:string,referer:string,agent:string,code:integer,size:integer
|
12
|
+
</filter>
|
13
|
+
|
14
|
+
<match azure-loganalytics.**>
|
15
|
+
@type azure-loganalytics
|
16
|
+
customer_id CUSTOMER_ID # Customer ID aka WorkspaceID String
|
17
|
+
shared_key KEY_STRING # The primary or the secondary Connected Sources client authentication key
|
18
|
+
log_type EVENT_TYPE_NAME # The name of the event type. ex) ApacheAccessLog
|
19
|
+
add_time_field true
|
20
|
+
time_field_name mytime
|
21
|
+
time_format %s
|
22
|
+
localtime true
|
23
|
+
add_tag_field true
|
24
|
+
tag_field_name mytag
|
25
|
+
</match>
|
@@ -20,7 +20,7 @@ Gem::Specification.new do |gem|
|
|
20
20
|
|
21
21
|
gem.add_dependency "fluentd", [">= 0.14.15", "< 2"]
|
22
22
|
gem.add_dependency "rest-client"
|
23
|
-
gem.add_dependency "azure-loganalytics-datacollector-api", [">= 0.1.
|
23
|
+
gem.add_dependency "azure-loganalytics-datacollector-api", [">= 0.1.5"]
|
24
24
|
gem.add_development_dependency "bundler", "~> 1.11"
|
25
25
|
gem.add_development_dependency "rake", "~> 10.0"
|
26
26
|
gem.add_development_dependency "test-unit"
|
Binary file
|
@@ -16,6 +16,8 @@ module Fluent::Plugin
|
|
16
16
|
:desc => "Your Operations Management Suite workspace ID"
|
17
17
|
config_param :shared_key, :string, :secret => true,
|
18
18
|
:desc => "The primary or the secondary Connected Sources client authentication key"
|
19
|
+
config_param :endpoint, :string, :default =>'ods.opinsights.azure.com',
|
20
|
+
:desc => "The service endpoint"
|
19
21
|
config_param :log_type, :string,
|
20
22
|
:desc => "The name of the event type that is being submitted to Log Analytics. log_type only alpha characters"
|
21
23
|
config_param :time_generated_field, :string, :default => '',
|
@@ -59,7 +61,7 @@ module Fluent::Plugin
|
|
59
61
|
def start
|
60
62
|
super
|
61
63
|
# start
|
62
|
-
@client=Azure::Loganalytics::Datacollectorapi::Client::new(@customer_id,@shared_key)
|
64
|
+
@client=Azure::Loganalytics::Datacollectorapi::Client::new(@customer_id,@shared_key,@endpoint)
|
63
65
|
end
|
64
66
|
|
65
67
|
def shutdown
|
@@ -22,18 +22,16 @@ class AzureLogAnalyticsOutputTest < Test::Unit::TestCase
|
|
22
22
|
|
23
23
|
def test_configure
|
24
24
|
d = create_driver
|
25
|
-
assert_equal '<Customer ID aka WorkspaceID String>', d.instance.customer_id
|
26
|
-
assert_equal '<Primary Key String>', d.instance.shared_key
|
27
25
|
assert_equal 'ApacheAccessLog', d.instance.log_type
|
28
|
-
|
29
|
-
|
30
|
-
|
26
|
+
assert_equal true, d.instance.add_time_field
|
27
|
+
assert_equal true, d.instance.localtime
|
28
|
+
assert_equal true, d.instance.add_tag_field
|
31
29
|
assert_equal 'tag', d.instance.tag_field_name
|
32
30
|
end
|
33
31
|
|
34
32
|
def test_format
|
35
33
|
d = create_driver
|
36
|
-
time = event_time("
|
34
|
+
time = event_time("2017-11-24 01:14:15 UTC")
|
37
35
|
d.run(default_tag: 'test') do
|
38
36
|
d.feed(time, {"a"=>1})
|
39
37
|
d.feed(time, {"a"=>2})
|
@@ -57,9 +55,9 @@ class AzureLogAnalyticsOutputTest < Test::Unit::TestCase
|
|
57
55
|
d.feed(
|
58
56
|
time,
|
59
57
|
{
|
60
|
-
:Log_ID => "
|
61
|
-
:date => "
|
62
|
-
:processing_time =>
|
58
|
+
:Log_ID => "5cdad72a-c848-4df0-8aaa-ffe033e75d57",
|
59
|
+
:date => "2017-11-24 01:44:32 JST",
|
60
|
+
:processing_time => 372,
|
63
61
|
:remote => "101.202.74.59",
|
64
62
|
:user => "-",
|
65
63
|
:method => "GET / HTTP/1.1",
|
@@ -67,15 +65,15 @@ class AzureLogAnalyticsOutputTest < Test::Unit::TestCase
|
|
67
65
|
:size => "-",
|
68
66
|
:referer => "-",
|
69
67
|
:agent => "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.7; rv:27.0) Gecko/20100101 Firefox/27.0",
|
70
|
-
:eventtime => "
|
68
|
+
:eventtime => "2017-11-24T01:44:32Z"
|
71
69
|
})
|
72
70
|
|
73
71
|
d.feed(
|
74
72
|
time,
|
75
73
|
{
|
76
|
-
:Log_ID => "
|
77
|
-
:date => "
|
78
|
-
:processing_time =>
|
74
|
+
:Log_ID => "7260iswa-8034-4cc3-uirtx-f068dd4cd659",
|
75
|
+
:date => "2017-11-24 01:45:14 JST",
|
76
|
+
:processing_time => 105,
|
79
77
|
:remote => "201.78.74.59",
|
80
78
|
:user => "-",
|
81
79
|
:method => "GET /manager/html HTTP/1.1",
|
@@ -83,7 +81,7 @@ class AzureLogAnalyticsOutputTest < Test::Unit::TestCase
|
|
83
81
|
:size => "-",
|
84
82
|
:referer => "-",
|
85
83
|
:agent => "Mozilla/5.0 (Windows NT 5.1; rv:5.0) Gecko/20100101 Firefox/5.0",
|
86
|
-
:eventtime => "
|
84
|
+
:eventtime => "2017-11-24T01:45:14Z"
|
87
85
|
})
|
88
86
|
end
|
89
87
|
data = d.events
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: fluent-plugin-azure-loganalytics
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 0.
|
4
|
+
version: 0.4.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Yoichi Kawasaki
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date:
|
11
|
+
date: 2019-06-06 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: fluentd
|
@@ -50,14 +50,14 @@ dependencies:
|
|
50
50
|
requirements:
|
51
51
|
- - ">="
|
52
52
|
- !ruby/object:Gem::Version
|
53
|
-
version: 0.1.
|
53
|
+
version: 0.1.5
|
54
54
|
type: :runtime
|
55
55
|
prerelease: false
|
56
56
|
version_requirements: !ruby/object:Gem::Requirement
|
57
57
|
requirements:
|
58
58
|
- - ">="
|
59
59
|
- !ruby/object:Gem::Version
|
60
|
-
version: 0.1.
|
60
|
+
version: 0.1.5
|
61
61
|
- !ruby/object:Gem::Dependency
|
62
62
|
name: bundler
|
63
63
|
requirement: !ruby/object:Gem::Requirement
|
@@ -113,10 +113,14 @@ files:
|
|
113
113
|
- README.md
|
114
114
|
- Rakefile
|
115
115
|
- VERSION
|
116
|
+
- docs/install-tdagent-and-the-plugin-on-rhel.md
|
116
117
|
- examples/fluent_1.conf
|
117
118
|
- examples/fluent_2.conf
|
119
|
+
- examples/fluent_csv.conf
|
120
|
+
- examples/fluent_typecast.conf
|
118
121
|
- fluent-plugin-azure-loganalytics.gemspec
|
119
122
|
- img/Azure-LogAnalytics-Fluentd.png
|
123
|
+
- img/Azure-LogAnalytics-Output-Image-2.png
|
120
124
|
- img/Azure-LogAnalytics-Output-Image.png
|
121
125
|
- lib/fluent/plugin/out_azure-loganalytics.rb
|
122
126
|
- test/helper.rb
|
@@ -140,8 +144,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
|
|
140
144
|
- !ruby/object:Gem::Version
|
141
145
|
version: '0'
|
142
146
|
requirements: []
|
143
|
-
|
144
|
-
rubygems_version: 2.5.2
|
147
|
+
rubygems_version: 3.0.3
|
145
148
|
signing_key:
|
146
149
|
specification_version: 4
|
147
150
|
summary: Azure Log Analytics output plugin for Fluentd
|