Find the best freelance Apache Spark jobs over 10 jobs for your full-time, part time or work from home opportunity and work with top rated clients on the top growing & trusted hiring platform connecting savvy businesses and professional freelancers.
Description:We have a running environment with MySQL db and we are starting to ingest more data than the db can handle, so we are looking for alternatives to architect Hadoop/Spark environment to offload most of that data into a Hadoop cluster.Activities:- Analyze our db and find the data that can be sent to Hadoop- Build a Hado...read more
Looking for an experienced coder, who can write code efficiently in Java and Scala for Apache Spark Streaming. He should also have a strong knowledge of kafka and should be able to create a data pipeline between kafka and spark for huge size of streaming data.
We are looking for freelancers who can create a crash course on Apache Spark. This will be for a course with 5 hours of content.Here are some other qualifications applicants will need:- Good English speaking and presenting skills - HD screencasts - Screen recording experience - Type the code out from scratch - A good microphoneP...read more
I need to convert JSON, Avro or other row-based format files in S3 into Parquet columnar store formats using an AWS service like EMR or Glue.I already have code that converts JSON to parquet using Python but the process is very manual, accounting for NULL values in the JSON elements by looking at each and every field/column and...read more
I am looking for a Hadoop developer with Java and scala experience. Looking for someone who have good experience and can take the pressure. This is a full time support and looking for someone who can work new york standard time from Mon thru Friday from 10 to 5 eastern time. Please don't waste my time and your time, if you are...read more
Looking for experienced developer on spark and scala with coding knowledge and implementation towards UDFs, UDTFs, UDAFs, actions, transformation, parallize on flatfiles, DBs, Hive tables and impact on memory optimization
Hi, what I need is the following: - I need a helm script (kubernetes) which 'installs' the latest spark version as yarn cluster on kubernetes - This setup should provide Zeppelin - All of this has to work with sparks latest structured streaming framework and should be able subscribe a kafka topic - Confluent is used as kafka pro...read more