Remote Web Development Job In IT And Programming

Backend developer needed to build a custom data scraper & parser service & API

Find more Web Development remote jobs posted recently Worldwide

The goal of this project is to ingest, sanitize, and structure data from LinkedIn in order to receive a continuous stream of updated profiles that meet specific criteria. The objective is to both obtain an initial dataset that is in a machine-readable format (CSV, XML, etc.) but then also to provide updated versions when a change is found.

Preliminary research into data access has shown that some information is available via:
The LinkedIn API
Query parameters in the URL that can be reverse-engineered to map to different values
Structured data accessible via a profile pages markup
Unstructured information
Different search products offered by LinkedIn (the free product, Recruiter, etc.).

Were looking to run searches for specific values (no current employment, number of years tenured in most recent position, etc.), keywords (practice area(s) selected from a specified list, etc.), and compound queries (attending a Top 50 ranked law school, Tier 1 law schools, etc.).

The types of queries were looking to track for a change are similar to:
When an individual from a specified list of companies leaves their current employer and changes to a new employment status that meets specific criteria.
When someone we currently track in our CRM switches jobs.
Etc.

We will be developing a separate rules engine that will provide the parameters, bounds, and frequency of updates for any given ongoing search. While the structure of these requests is not yet defined, the tool should be built so that a query can be carried out based on a set of parameters (JSON, etc.).

The tool should be able to search on behalf of specific employees who have granted the tool access to their accounts in aggregate.

Results and updates should be delivered every week. This check could be done via polling only for deltas, comparing two entire data sets, etc. The data should have a flag and a unique identifier to signal whether each result is new or an update.

Each batch of results should also have relevant timestamps to reflect both the time of change and the time the change was crawled.

As an output, we will need all of the data in a single file in a machine-readable format (CSV, XML, etc.).

Additionally, providing the data via a documented API that delivers the output in XML/JSON could be part of the initial scope or another phase.
About the recuiter
Member since Mar 14, 2020
Peter J Krenz
from Lhuntshi, Bhutan

Skills & Expertise Required

Data Scraping Web Scraping 

Open for hiringApply before - Aug 30, 2024

Work from Anywhere

40 hrs / week

Hourly Type

Remote Job

$17.25

Cost

Offer to work on this project closes in 58 days!
Are you interested in this Opportunity?

Looking for help? Checkout our video tutorial
How to search and apply for jobs

How to apply? Do you have more questions about the Job?
See frequently asked questions

Similar Projects

Data Extraction from Public Database into Daily Report

We are looking for a freelancer to build us an excel spreadsheet that can pull data from Public Database into Daily Report.

We would like the excel spreadsheet to lists the title, opportunity type (solicitation, sources sought, etc.), docum...read more

Need Zip Code Data for USA (Zip Code of Universities, Malls, Golf Courses, etc)

I want to be able to have zip codes of anything that can be considered a demographic.

For example which Zip Codes are primarly Colleges/Universities
Which Zip Codes have Golf Courses in them.
Which Zip Codes have Malls around themread more

Python scraping script fix with selenium

You need to have access to fanduel, which has a location striction. I will provide the account.
The script is written in Python and using Selenium.

Chinese undergrad students application and enrolled numbers in top 60 Universities for 20 years

1. collecting authentic data of Chinese students application and enrolled numbers in US Universities.( the best place you can search is from common app. org, but i dont know if they share data)

2. collecting huge data sets of Chinese studen...read more