I have an existing Map Reduce Job implemented in Java. Normally I should just run hadoop jar args... with appropriate arguments, and with HADOOP_CLASSPATH set to load supporting jar files. I'm new to DSE and I have no idea how to do this here. My first blind try was to run dse hadoop jar... but this didn't work. Please help me on this!
In an effort to consolidate free help offered for our products we have decided to move these forums to a more widely used forum. Please use one of the following queries (or any combination):
- Cassandra: tag search or plain text search
- DataStax Enterprise: tag search or plain text search
- DataStax OpsCenter: tag search or plain text search
How to run DSE hadoop with existing Map Reduce Job(2 posts) (2 voices)
DSE uses CassandraFS to replace HDFS, so you need to add your data to CassandraFS.
You can also store your data to Cassandra CF, then use job.setInputFormatClass(ColumnFamilyInputFormat.class);
as your input and out format. check this link http://www.datastax.com/dev/blog/whats-new-cassandra-07-hadoop-output-cassandra
You may need re-compile your mapreduce code with DSE bundled haddoop jar file
run it at DSE install folder as bin/dse hadoop jar ......