Algorithm in hadoop and spark (2 versions)

进行中 已发布的 4 年前 货到付款
进行中 货到付款

Hi!

Looking for developer to help me with porting algorithm (details via email) to hadoop and to spark. More or less idea is to compute formula similar to Pearson correlation (formula in files as image).

Payment is open for discussion.

some clarification needed:

one csv file with structure of item_id, item_name, item_group

second one: user_id, item_id, rating

the algorithm:

for upfront stated user_id (user_0)

divides input from second file by item_group

for each group

calculates Pearson correlation between user_0 and rest of users (considering ratings only from this specific group)

produces top N most correlated users (N can be as const in the program)

from that top N list calculates averages for each item rating and returns top M items (M can be const as well)

output: topM items for each group

Hadoop Java Spark

项目ID: #22190052

关于项目

7个方案 远程项目 活跃的4 年前

授予:

vmrraj2020

This is Mohanraj V. I have completed Master of Computer Applications.I have 6 years of IT experience in Data scientist and Hadoop development including Python,HDFS,HIVE, PySpark ,HBase and Experienced in SQL and PL/SQ 更多

$240 USD 在1天内
(0条评论)
0.0

有7名威客正在参与此工作的竞标,均价$198/小时

rnaushad

Hi, I have around 15 years of experience in the java stack and have been working on the big data stack for the past 2 years. I have a very good handle on Spark jobs and have written some very complex jobs in spark and 更多

$111 USD 在2天内
(9条评论)
3.1
jonyahmed105

Hi I cn handle it. THANKS

$250 USD 在3天内
(5条评论)
4.1
TechnicalGeeks

Hi, I am a bigdata developer and a module lead in reputed MNC.i an into the IT industry for more then 12 years. I have tonnes of experience in developing projects using Java,Apache Spark,Hive,Kafka,Scoop,Pig,Scala,aws 更多

$250 USD 在5天内
(1条评论)
1.7
anviverma111

HI I am experienced in Java Hadoop Spark etc I can start right now but i have few doubts and questions lets have a quick chat and get it started waiting for your reply

$140 USD 在7天内
(0条评论)
0.0
AnupKumarSaha30

 Expertise in all components of Hadoop Ecosystem- Hive, Hue, Pig, Sqoop, HBase, Flume, Zookeeper, Oozie, Apache Flink and Apache Spark.  Responsible for writing MapReduce programs using Java.  Logical Implementation 更多

$120 USD 在7天内
(0条评论)
0.0
ehanson5

I am a lead data engineer and develop complex Spark applications on Hadoop on a daily basis. once I have full requirements I will deliver within 3 days

$277 USD 在3天内
(0条评论)
0.0