Comments (6)
dfs.support.broken.append
is not required for HDFS latest versions.
And your error Connection refused - connect(2)
means that fluentd failed to connect on TCP layer, not on WebHDFS protocol layer. It's not problem of hadoop versions.
You should check the connectivities between your fluentd (td-agent) host and all of your cluster nodes (namenode and all of datanodes).
from fluent-plugin-webhdfs.
Hi Tagamoris,
Thanks for your response. In terms of TCP-connectivity, i am able to perform telnet and nc, so i dont think tcp-connectivity is an issue.
Is there something i might have missed in terms of testing tcp-connectivity?
Thanks in advance
Alex
from fluent-plugin-webhdfs.
How do you check tcp connectivity (in command)?
Or, can you put any data from fluentd host with curl?
curl -v -XPUT --data @file.txt "http://your.namenode.local:50070/webhdfs/v1/file.txt?op=CREATE&user.name=hdfs"
# => redirected
curl -v -XPUT --data @file.txt REDIRECTED_URL
from fluent-plugin-webhdfs.
Hi Tagomoris,
Thank you so much for your reply. I tried those cURL commands to PUT a file via webhdfs.
Basically it reports that:
Connected to datanode.host.com (my-ip) port 50075 (#0)
PUT /webhdfs/v1/foo.txt?op=CREATE HTTP/1.1
User-Agent: curl/7.19.7 (x86_64-redhat-linux-gnu) libcurl/7.19.7 NSS/3.14.0.0 zlib/1.2.3 libidn/1.18 libssh2/1.4.2
Host: datanode.host.com:50075
Accept: /
Content-Length: 24
Content-Type: application/x-www-form-urlencoded< HTTP/1.1 400 Bad Request
< Cache-Control: no-cache
< Expires: Tue, 06 May 2014 13:51:10 GMT
< Date: Tue, 06 May 2014 13:51:10 GMT
< Pragma: no-cache
< Expires: Tue, 06 May 2014 13:51:10 GMT
< Date: Tue, 06 May 2014 13:51:10 GMT
< Pragma: no-cache
< Content-Type: application/json
< Transfer-Encoding: chunked
< Server: Jetty(6.1.26)
<
- Connection #0 to host datanode.host.com left intact
- Closing connection #0
{"RemoteException":{"exception":"IllegalArgumentException","javaClassName":"java.lang.IllegalArgumentException","message":"namenoderpcaddress is not specified."}}
So it is probably due to the dfs.namenode.servicerpc-address setting.
We tried it by specifying the port to 8022 and also tried the default value (empty), but both scenarios complain about the same error (namenoderpcaddress is not specified).
So are there certain gotcha's regarding HDFS settings that we need to be aware of apart from the ones mentioned here http://docs.fluentd.org/articles/out_webhdfs
Thanks in advance
Alex
from fluent-plugin-webhdfs.
Oops.... problem seems to be solved now.
It seems that if i create a directory beforehand by "hadoop fs" commands, with owner hdfs and mode 777, it actually wont let me write files to that directory.
However, if i dont create a directory beforehand, somehow it does allow me to put files in the specified paths in td-agent.conf
Thanks again for your help!
from fluent-plugin-webhdfs.
OK, I see. I'll close this issue.
from fluent-plugin-webhdfs.
Related Issues (20)
- Plugin lost file buffer when setting buffer type to file and split file by hourly HOT 5
- forward error error=#<Encoding::UndefinedConversionError: "\xE6" from ASCII-8BIT to UTF-8> error_class=Encoding::UndefinedConversionError HOT 7
- Output by tag like out_file (Feature Request) HOT 3
- 根据日志中的字段指定hdfs上的路径 HOT 5
- httpFS - Do not create file if it does not exist HOT 5
- part of data missing
- output_format , just simple HOT 12
- out_webhdfs.rb's class changed from TimeSlicedOutput to Output HOT 2
- the webhdfs doesn't work HOT 6
- Kerberos Keytab example HOT 5
- Can hdfs path use ${tag} or ${record} ? just like: /data/hdfs/${tag}.#{Socket.gethostname}.log HOT 1
- Error installing fluent-plugin-webhdfs HOT 3
- HA configuration performs incorrectly HOT 4
- When try to append to file more often then default timekey, exception happens
- kerberos_keytab not authorize HOT 2
- use chunk_id in path HOT 3
- httpFS - Can not create file when it does not exist
- Compression Snappy is not work. HOT 3
- systemctl reload td-agent.service is not working properly with @type webhdfs
- Operation category READ is not supported in state standby
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from fluent-plugin-webhdfs.