MAP REDUCE - HADOOP - SPARK - PYTHON: Need algorithms written to be used with my data-sets

Completed Posted 7 years ago Paid on delivery
Completed Paid on delivery

Hello,

First and foremost, the requirements of this project are not very technical.

PROJECT DESCRIPTION: I need someone who can get the job done without asking too many questions, someone who is proficient in using Hadoop, MAP Reduce, Spark and can code in python.

If you have the skills as mentioned above, then please continue reading.

You must build me the following:

- Implement an algorithm to obtain the first (n) most frequent words in a large text file using Hadoop Map reduce. (n) = is the data-set, which you will be provided.

- Implement an algorithm to obtain the (AVERAGE) of all (n) (MOST FREQUENT WORDS) that appear in the text file using Hadoop Map reduce.

- Implement an algorithm to obtain the first (n) most frequent words in a large text file using Hadoop Spark.

- Ability to use (NMON) to compare the performances of the algorithms.

- Implement a simple (MIN-HASHING) algorithm, this must be written in Hadoop Spark. This algorithm will be used to compare 5 files, where their signatures will be used by (JACCARD SIMILARITY METRIC).

- Testing all algorithms with other provided data-sets.

You must absolutely be able to do what has been mentioned in the project description.

Please do not BID if you are not capable of doing this project.

Thank you!

Big Data Sales Hadoop Map Reduce Python Spark

Project ID: #11916153

About the project

8 proposals Remote project Active 7 years ago

Awarded to:

iyersume

Hi, I am proficient in Python, PySpark, MapReduce and Data Mining algorithms. I would like the opportunity to work with you on this project. Please feel free to ping on chat to discuss more. I have a few questions e.g. More

$200 USD in 7 days
(39 Reviews)
5.4

8 freelancers are bidding on average $234 for this job

sumbali

Hi there, I have gone through the details and this is something I'm good at. I am an experienced independent academic writer with excellent research skills. I am known for my quality and commitment. Please go through m More

$250 USD in 3 days
(9 Reviews)
4.7
amitkumar0327

Hi, I am Amit. I have experience in MapReduce ,Spark,Java and Scala. I can write the alogithms for all the requirement in the project. Looking forward to work on this. Regards, Amit Kumar

$200 USD in 5 days
(14 Reviews)
4.2
rexzetsolutions

i am well versed in Hadoop, MAP Reduce, Spark, and have skills in python and djanjo, i read the project and i am capable t

$200 USD in 2 days
(6 Reviews)
3.2