Remote Data Mining And Management Job In Data Science And Analytics

Build out Advance Analytics and BI Platform

Find more Data Mining And Management remote jobs posted recently Worldwide

I will share my use case and am looking to build out Advance analytics and BI platform thats cost-effective yet viable (capable of working successfully) and scalable with shortliste candidates.

Current Use Case is to collect data from CRM service providers like SalesForce &
Microsoft CRM 365 and then transform into meaningful data by logically joining
different entities received and then persisting it into Data Warehouse.
Business Intelligence Dashboards will be developed and integrate it with Data
Warehouse and perform analysis using analytical queries (joining, grouping, sorting
etc.

Along with this, the Advanced Analytics Platform will be developed in which Data Science
The team will perform basic analysis first and Then build & train their Machine Learning
Models on top of it for Predictive Analytics & Recommendation Engines.

Create below modules -

1. Data Sources Management
Using this Module, User will be able to configure their Data Sources from which Data
needs to be collected.
Once User configure the Data Sources, Data Ingestion Job will be submitted to Apache
Gobblin and Gobblin will start collecting data from Data Sources.

2. Real-Time Analytics
Once Ingested Data is available on Kafka Streams, Structured Spark Streaming will be
used to process & transform it in a distributed way and write that to MariaDB Column
Store.

3. Data Lake
Data Lake will be required as Lot of Data will be ingested and Data Warehouse will be
having cleaned & transformed version of data .
All Data collected from CRM Data Sources will be coming in JSON format . So Apache
Gobblin will convert the json data into parquet format before loading in Data Lake for
better I/O & Low Latency Reads

4. Data Processing
For Data Processing, Apache Spark will be used which is distributed data processing
engine and Data Processing Jobs will be scheduled using Apache Airflow and it will
read latest data from Data Lake and apply required transformations and then persist the
data to Data Warehouse.

5. Data Warehousing
For Data Warehouse, Hive on Minio will be used and File Format will be Parquet. The hive
will act as a MetaStore and Schemas will be defined in it for various tables and Tables
will be pointing to their corresponding Minio Storage Location.

Business Intelligence
Both Querying Engines i.e. Spark sql on Hive and MariaDb ColumnStore supports JDBC .
So , Any BI Tool can connect to them using Standard JDBC Connections and execute
analytics queries and create various charts/graphs.
About the recuiter
Member since Mar 14, 2020
Shamsaagazarzoo
from Bacs-Kiskun, Hungary

Skills & Expertise Required

Apache Hive Apache Kafka Apache Spark Backend Rest API Minio 

Open for hiringApply before - Sep 9, 2024

Work from Anywhere

40 hrs / week

Fixed Type

Remote Job

$19,141.05

Cost

Offer to work on this project closes in 53 days!
Are you interested in this Opportunity?

Looking for help? Checkout our video tutorial
How to search and apply for jobs

How to apply? Do you have more questions about the Job?
See frequently asked questions

Similar Projects

Train me on hadoop ecosystem.

1. Sourcing data from multiple system into different file formats.
2. Ingesting the extracted data into hadoop cluster (csv, xml, json, fixed width etc)
3. Using mapreduce, pig to transform the data and load it into mpp.
4. Stitch all abo...read more

Approximate synchronous Kafka copy

The task is to write a distributed approximately synchronous copy procedure for Kafka partitions. The entire applications is envisioned to be provisioned as a helm chart and the application is supposed to be os.

I will share some files with...read more

Small Task on Python & Spark (PySpark)

HI,
I need a small less than 1 hour task to be done using Pyspark.
Add some spark code to existing python code.

Converting JSON or Avro files to Parquet

I need to convert JSON, Avro or other row-based format files in S3 into Parquet columnar store formats using an AWS service like EMR or Glue.

I already have code that converts JSON to parquet using Python but the process is very manual, acco...read more