The output from the Hadoop MapReduce wordcount example is sorted by the key. So the output should be in alphabetical order.
With Hadoop you can create your own key objects that implement the WritableComparable
interface allowing you to override the compareTo
method. This allows you to control the sort order.
To create an output that is sorted by the number of occurances you would probably have to add another MapReduce job to process the output from the first as you have said. This second job would be very simple, maybe not even requiring a reduce phase. You would just need to implement your own Writable
key object to wrap the word and its frequency. A custom writable looks something like this:
public class MyWritableComparable implements WritableComparable {
// Some data
private int counter;
private long timestamp;
public void write(DataOutput out) throws IOException {
out.writeInt(counter);
out.writeLong(timestamp);
}
public void readFields(DataInput in) throws IOException {
counter = in.readInt();
timestamp = in.readLong();
}
public int compareTo(MyWritableComparable w) {
int thisValue = this.value;
int thatValue = ((IntWritable)o).value;
return (thisValue < thatValue ? -1 : (thisValue==thatValue ? 0 : 1));
}
}
I grabbed this example from here.
You should probably override hashCode
, equals
and toString
as well.