Data Extraction Freelance Jobs

Find Data Extraction jobs online. Work from home and earn money online.
Signup for free to get access to all filter attributes and instant notifications when new jobs are posted.
Setup filter



Get access to over 30+ filter attributes, setup instant notifications, integrate with your CRM and marketing tools, and more.
Start free trial
376 projects published for past 72 hours.
Job Title Budget
Mathematical Symbols Detection in PDF and Images
~147 - 442 USD 2 minutes ago
Client Rank - Medium

Payment method verified
2 open job
no reviews
Registered at: 22/06/2015
IN India
Medium
I'm seeking a professional who can help me detect mathematical symbols in both PDF and image files. The results should be delivered in a JSON format.

Ideal skills and experience for the job:
- Proficiency in handling PDF and image files
- Strong understanding of mathematical symbols
- Experience in data processing and result formatting in JSON
- Attention to detail to ensure no symbols are missed
- Ability to deliver clean and well-structured output
- Proficiency in using OCR (Optical Character Recognition) technology for symbol detection.
- Background in machine learning and AI for improving detection accuracy.
- Familiarity with LaTeX for understanding and formatting mathematical symbols.
- Expertise in computer vision to accurately recognize mathematical symbols.
- Experience in writing Python scripts for file manipulation and data extraction.
- Knowledge of JSON schemas to ensure well-structured output data.
- Knowledge of natural language processing (NLP) techniques for text and symbol parsing.
- Experience in integrating with APIs for advanced processing and verification.
- Skills in optimizing detection algorithms for speed and accuracy.
- Experience in verifying and validating the accuracy of detected symbols.

Please propose how you would approach this task, including any tools or methods you plan to use.

Accuracy level required for this project is high. Special focus should be given to detecting advanced mathematical symbols such as integrals, sums, and limits. The system should detect symbols within a few seconds. The primary documents for this project will be scanned PDFs.

Skills: PHP, Python, Matlab and Mathematica, LaTeX, Machine Learning (ML)
Fixed budget: 12,500 - 37,500 INR
2 minutes ago
  • Websites, IT & Software, Writing & Content, Design, Media & Architecture, Engineering & Science, Python, LaTeX, Matlab and Mathematica, Machine Learning (ML)
Website developer Front and Back end
not specified 28 minutes ago
Client Rank - Risky

Payment method not verified
1 open job
SG Singapore
Risky
Required Connects: 8
Hi there, looking for a web developer for both front & backend. Includes some good sense of UX & UI. We are basically looking for a full-time to work with us by developing websites both internally and externally. If you are interested. Please do reply back and hope to hear from you soon.
Skills: HTML, CSS, JavaScript, Python, React, Vue.js, Redux, MongoDB, Postman, PostgreSQL, Data Scraping, Data Extraction, Web Scraping, Web Crawling
Budget: not specified
28 minutes ago
  • Web, Mobile & Software Dev, Web Development
Website developer Front and Back end
not specified 28 minutes ago
Client Rank - Risky

Payment method not verified
1 open job
SG Singapore
Risky
Required Connects: 8
Hi there, looking for a web developer for both front & backend. Includes some good sense of UX & UI. We are basically looking for a full-time to work with us by developing websites both internally and externally. If you are interested. Please do reply back and hope to hear from you soon.
Skills: HTML, CSS, JavaScript, Python, React, Vue.js, Redux, MongoDB, Postman, PostgreSQL, Data Scraping, Data Extraction, Web Scraping, Web Crawling
Budget: not specified
28 minutes ago
  • Web, Mobile & Software Dev, Web Development
Webscraping Specialist with SEMRush and API Development Skills
20 - 40 USD / hr
42 minutes ago
Client Rank - Medium

Payment method verified
1 jobs posted
1 open job
no reviews
Registered at: 25/09/2024
GB United Kingdom
Medium
Required Connects: 15
We are seeking an experienced webscraping specialist proficient in Python, specifically using SEMRush for data extraction. The ideal candidate should have a solid background in API development to ensure seamless integration and data handling. Familiarity with AWS S3 for data storage and Snowflake for data validation is a plus. If you have a keen eye for detail and a passion for data-driven solutions, we want to hear from you!
Skills: API, PHP, Data Scraping, JavaScript, Python
Hourly rate: 20 - 40 USD
42 minutes ago
  • Web, Mobile & Software Dev, Scripts & Utilities
Web Scraping for Company Data Extraction
50 USD 55 minutes ago
Client Rank - Medium

Payment method verified
$42 total spent
2 hires
5 jobs posted
40% hire rate, 1 open job
5.00 of 1 reviews
Registered at: 30/05/2020
IN India
Medium
Required Connects: 11
We are seeking a skilled web scraper to extract a list of companies along with their city, country, company name, Emails, phone numbers and website from a specified website. The ideal candidate should have experience with web scraping tools and techniques, and be able to deliver clean, organized data in a timely manner. Familiarity with handling various web formats and a keen eye for detail are essential. If you have expertise in this area and can provide a reliable solution, please apply!
Skills: Data Scraping, Data Mining, Data Extraction, Scrapy
Fixed budget: 50 USD
55 minutes ago
  • Data Science & Analytics, Data Extraction/ETL
Enrich linkedin company URL
30 USD 1 hour ago
Client Rank - Medium

Payment method verified
$127 total spent
2 hires , 2 active
1 open job
4.86 /hr avg hourly rate paid
23 hours
no reviews
Registered at: 04/11/2024
FR France
Medium
Required Connects: 8
Hi Muhammad,

I have a new need, like last time. I need to find the linkedin company URLs from the list in PJ.
Skills: Data Entry, Lead Generation, B2B Lead Generation, Data Mining, Data Collection, Data Extraction, Market Research, Web Scraping, Social Media Lead Generation, LinkedIn Sales Navigator, Data Cleaning, Email List, Online Research, Communications, Prospect List
Fixed budget: 30 USD
1 hour ago
  • Admin Support, Data Entry & Transcription Services
Senior Data Engineer (Snowflake/ETL)
22,000 USD 1 hour ago
Client Rank - Medium

Payment method verified
71 jobs posted
1% hire rate, 8 open job
no reviews
Registered at: 09/05/2024
IN India
Medium
Required Connects: 10
Senior Data Engineer (Snowflake/ETL)
Experience: 8+ Years
What makes you a great fit:
 You have 8+ years of extensive development experience using snowflake or
similar data warehouse technology
 You have working experience with dbt and other technologies of the modern
datastack, such as Snowflake, Apache Airflow, Fivetran, Looker, AWS, git.
 You have experience in agile processes, such as SCRUM
 You have extensive experience in writing advanced SQL statements and
performance tuning them
 You have experience in Data Ingestion techniques using custom or SAAS tool
like fivetran
 You have experience in data modelling and can optimise existing/new data
models
 You have experience in data mining, data warehouse solutions, and ETL, and
using databases in a business environment with large-scale, complex datasets
 You having experience architecting analytical databases (in Data Mesh
architecture) is added advantage
 You have experience working in agile cross-functional delivery team
 You have high development standards, especially for code quality, code reviews,
unit testing, continuous integration and deployment
As a Analytics Engineer you’ll be:
 Developing end to end ETL/ELT Pipeline working with Data Analysts of business
Function.
 Designing, developing, and implementing scalable, automated processes for
data extraction, processing, and analysis in a Data Mesh architecture
 Mentoring other Junior Engineers in the Team
 Be a “go-to” expert for data technologies and solutions
 Ability to provide on the ground troubleshooting and diagnosis to architecture
and design challenges
 Troubleshooting and resolving technical issues as they arise
 Looking for ways of improving both what and how data pipelines are delivered
by the department
 Translating business requirements into technical requirements, such as entities
that need to be modelled, DBT models that need to be build, timings, tests and
reports
 Owning the delivery of data models and reports end to end
 Perform exploratory data analysis in order to identify data quality issues early in
the process and implement tests to ensure prevent them in the future
 Working with Data Analysts to ensure that all data feeds are optimised and
available at the required times. This can include Change Capture, Change Data
Control and other “delta loading” approaches
 Discovering, transforming, testing, deploying and documenting data sources
 Applying, help defining, and championing data warehouse governance: data
quality, testing, coding best practices, and peer review
 Building Looker Dashboard for use cases if required
Skills: Looker, AWS,, ETL
Fixed budget: 22,000 USD
1 hour ago
  • Data Science & Analytics, Data Mining & Management
Data Engineer with Snowflake and Looker
25,000 USD 1 hour ago
Client Rank - Medium

Payment method verified
71 jobs posted
1% hire rate, 8 open job
no reviews
Registered at: 09/05/2024
IN India
Medium
Required Connects: 10
Senior Data Engineer (Snowflake/ETL)
Experience: 8+ Years
• What makes you a great fit:
• You have 5+ years of extensive development experience using Snowflake or
• similar data warehouse technology
• You have working experience with dbt and other technologies of the modern
• datastack, such as Snowflake, Apache Airflow, Fivetran, Looker, AWS, git.
• You have experience in agile processes, such as SCRUM
• You have extensive experience in writing advanced SQL statements and
• performance tuning them
• You have experience in Data Ingestion techniques using custom or SAAS tool
• like fivetran
• You have experience in data modelling and can optimise existing/new data
• models
• You have experience in data mining, data warehouse solutions, and ETL, and
• using databases in a business environment with large-scale, complex datasets
• You having experience architecting analytical databases (in Data Mesh
• architecture) is added advantage
• You have experience working in agile cross-functional delivery team
• You have high development standards, especially for code quality, code reviews,
• unit testing, continuous integration and deployment
• As a Analytics Engineer you’ll be:
• Developing end to end ETL/ELT Pipeline working with Data Analysts of business
• Function.
• Designing, developing, and implementing scalable, automated processes for
• data extraction, processing, and analysis in a Data Mesh architecture
• Mentoring other Junior Engineers in the Team
• Be a “go-to” expert for data technologies and solutions
• Ability to provide on the ground troubleshooting and diagnosis to architecture
• and design challenges
• Troubleshooting and resolving technical issues as they arise
• Looking for ways of improving both what and how data pipelines are delivered
• by the department
• Translating business requirements into technical requirements, such as entities
• that need to be modelled, DBT models that need to be build, timings, tests and
• reports
• Owning the delivery of data models and reports end to end
• Perform exploratory data analysis in order to identify data quality issues early in
• the process and implement tests to ensure prevent them in the future
• Working with Data Analysts to ensure that all data feeds are optimised and
• available at the required times. This can include Change Capture, Change Data
• Control and other “delta loading” approaches
• Discovering, transforming, testing, deploying and documenting data sources
• Applying, help defining, and championing data warehouse governance: data
• quality, testing, coding best practises, and peer review
• Building Looker Dashboard for use cases if required
Skills: Looker, AWS,, Git, SQL CLR, ETL
Fixed budget: 25,000 USD
1 hour ago
  • Data Science & Analytics, Data Mining & Management
Data Collection & Email List Building
10 USD 1 hour ago
Client Rank - Excellent

Payment method verified
$4'840 total spent
154 hires , 7 active
158 jobs posted
97% hire rate, 72 open job
4.82 /hr avg hourly rate paid
77 hours
4.99 of 207 reviews
Registered at: 05/09/2024
US United States
Excellent
Required Connects: 8
I need someone to Build Email list Target Companys Interested please apply.
Skills: Data Entry, Lead Generation, Lead Generation Analysis, Online Research, Data Scraping, Data Extraction, Data Analysis, Real Estate, Data Analytics, Data Collection
Fixed budget: 10 USD
1 hour ago
  • Admin Support, Data Entry & Transcription Services
Scrape city ordinance website
20 - 35 USD / hr
2 hours ago
Client Rank - Good

Payment method verified
$7'482 total spent
9 hires
7 jobs posted
100% hire rate, 1 open job
32.27 /hr avg hourly rate paid
210 hours
5.00 of 6 reviews
Registered at: 25/12/2021
US United States
Good
Required Connects: 14
We have a requirement to scrape data from a website that hosts city ordinance / municipal data.

Because the documents are so large, we only wish to extract a certain chapter regarding a specific type of regulation from each city.

The next challenge we have is that the chapter we're looking for differs between cities (both in name and content), and to identify the correct chapter we need to query the contents for certain keywords

This job requires someone who has experience in web scraping APIs, and knows how to query data to find whats needed. Ideally we'd like the deliverable as a Python file.

The website in question is attached in the TXT file to this job
Skills: Data Scraping, Data Mining, Python, Web Crawling, Data Extraction, Scrapy
Hourly rate: 20 - 35 USD
2 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Scrape data from adult sites
100 USD 2 hours ago
Client Rank - Medium

Payment method verified
$728 total spent
3 hires , 2 active
3 jobs posted
100% hire rate, 1 open job
13.96 /hr avg hourly rate paid
41 hours
5.00 of 1 reviews
Registered at: 05/03/2024
IN India
Medium
Required Connects: 14
I would need the following data only from cities in India - Female and Shemale profile ) from 2 websites

1. Phone
2.Username
3.Location
4.Gender
5.Email address
6.Profile URL
7. Review & Rating ( There would be multiple on a page want would want all of them )

Would need the need data in an Excel format
Skills: Data Scraping, Data Entry, Data Extraction, Data Mining, Microsoft Excel
Fixed budget: 100 USD
2 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Scrape data from PropWire using Excel VBA and bypassing Captcha.
100 USD 3 hours ago
Client Rank - Good

Payment method verified
$1'092 total spent
9 hires , 1 active
12 jobs posted
75% hire rate, 1 open job
65.19 /hr avg hourly rate paid
13 hours
5.00 of 7 reviews
Registered at: 19/06/2021
US United States
Good
Required Connects: 14
I need an Excel VBA script that that will open PropWire paste the string variable and collect the data PropWire returns using the clipboard. The program should capture EVERYTHING that is on the screen, yes garbage data and all, and my program will process what has been collected. The subroutine will start with this line...

Sub CapturePropWireOnlineData(sv as String)

I can provide as much sample data you require.
Please Note: I don't like Selenium because I have to update each computer but if that is the only way, I will accept Selenium code.
Also, I do not want the program written in Python for the same reason and I will not accept offers for Python scripting.

Please note that one person gave up in his attempt to complete this project.
Additional/Future projects: Clicking on the Propwire links, Owner, Comps and History to collect the data from each page using the clipboard.

Let me emphasize, Python is not an option, this needs to done using selenium. Also the speed it runs is not a priority, getting around the automation detection is.
Skills: Microsoft Excel, Visual Basic for Applications, Data Scraping, Data Mining, Data Extraction, Automation, Scripting
Fixed budget: 100 USD
3 hours ago
  • Web, Mobile & Software Dev, Scripts & Utilities
Automation Workflows for Data
25 - 50 USD / hr
3 hours ago
Client Rank - Risky

Payment method not verified
2 open job
no reviews
Registered at: 25/11/2024
GB United Kingdom
Risky
Check the attached document for a better overview of the project info!

We are seeking an experienced UiPath freelancer to develop workflows for extracting data from Excel, PDF, Word, and Email files, converting it into structured formats (e.g., JSON, CSV, XML). This project serves as the foundation for a broader workflow automation initiative. If successful, there will be opportunities for ongoing development work to expand and enhance the system.
Provided Tools
We will supply:
UiPath Pro License: For workflow development.
UiPath Document Understanding: For advanced OCR and document processing.

Proposal Submission Guidelines
Freelancers are invited to submit proposals incorporating an hourly rate for the project. Along with your hourly rate, please provide the following details for each file type:
Milestone Plan 
Outline the steps for developing and testing workflows for each file type.
Provide a timeline and estimated effort in hours for each milestone.


Existing Tools and Solutions
Highlight any prebuilt workflows, templates, or components you plan to use.
The use of prebuilt tools or those sourced from the UiPath Marketplace will be considered positively and demonstrate efficiency.
Approach
Detail your methodology for each milestone, including: 
Tools and components (e.g., UiPath Document Understanding, OCR engines, plugins).
Error-handling mechanisms to ensure robustness and scalability.
Plans to adapt or customize any prebuilt tools.
Timeline
Specify the time required to: 
Develop or customize workflows for each file type.
Test and validate workflows using sample files provided.
Hourly Rate
Indicate your proposed hourly rate.
Provide a rough breakdown of total expected hours for each workflow type.

Evaluation Criteria
Proposals will be evaluated based on:
Hourly Rate and Estimated Time: Competitiveness and feasibility of proposed hours.
Optimisation: Use of prebuilt tools, templates, or UiPath Marketplace components.
Scalability: Workflows should handle large volumes of files efficiently.
Robustness: Strong error-handling and logging mechanisms.
Documentation: Clear and user-friendly workflow guides.

Submission Deadline
Task Deadline
Proposal Due 17.12.24
Project Start 06.01.25

Deliverables
UiPath Workflows: Functional workflows for each file type (PDF, Excel, Word, and Email).
Documentation: Clear user guides, workflow diagrams, and troubleshooting notes.
Test Cases: Validation results using sample files provided.

Ongoing Development Opportunities
If successful in this project, the freelancer may be considered for:
Expanding workflows to additional file types (e.g., CAD, GIS).
Enhancing workflows for advanced error handling, compliance mechanisms, and scalability.
Integrating anonymisation processes to ensure GDPR compliance.

By incorporating an hourly rate, this approach ensures flexibility while allowing us to evaluate proposals based on time efficiency, experience, and cost. We look forward to receiving your proposals!

This version emphasizes the hourly rate approach and invites freelancers to provide detailed estimates and methods for each milestone while allowing flexibility for negotiation.

Skills: Machine Learning (ML), Artificial Intelligence, Data Scraping, Data Extraction, Deep Learning
Hourly rate: 25 - 50 USD
3 hours ago
  • Websites, IT & Software, Data Entry & Admin, Engineering & Science, Artificial Intelligence, Data Scraping, Data Extraction, Machine Learning (ML), Deep Learning
Automation Workflows for Data
25 - 50 USD / hr
3 hours ago
Client Rank - Risky

Payment method not verified
2 open job
no reviews
Registered at: 25/11/2024
GB United Kingdom
Risky
Check the attached document for a better overview of the project info!

We are seeking an experienced UiPath freelancer to develop workflows for extracting data from Excel, PDF, Word, and Email files, converting it into structured formats (e.g., JSON, CSV, XML). This project serves as the foundation for a broader workflow automation initiative. If successful, there will be opportunities for ongoing development work to expand and enhance the system.
Provided Tools
We will supply:
UiPath Pro License: For workflow development.
UiPath Document Understanding: For advanced OCR and document processing.

Proposal Submission Guidelines
Freelancers are invited to submit proposals incorporating an hourly rate for the project. Along with your hourly rate, please provide the following details for each file type:
Milestone Plan 
Outline the steps for developing and testing workflows for each file type.
Provide a timeline and estimated effort in hours for each milestone.


Existing Tools and Solutions
Highlight any prebuilt workflows, templates, or components you plan to use.
The use of prebuilt tools or those sourced from the UiPath Marketplace will be considered positively and demonstrate efficiency.
Approach
Detail your methodology for each milestone, including: 
Tools and components (e.g., UiPath Document Understanding, OCR engines, plugins).
Error-handling mechanisms to ensure robustness and scalability.
Plans to adapt or customize any prebuilt tools.
Timeline
Specify the time required to: 
Develop or customize workflows for each file type.
Test and validate workflows using sample files provided.
Hourly Rate
Indicate your proposed hourly rate.
Provide a rough breakdown of total expected hours for each workflow type.

Evaluation Criteria
Proposals will be evaluated based on:
Hourly Rate and Estimated Time: Competitiveness and feasibility of proposed hours.
Optimisation: Use of prebuilt tools, templates, or UiPath Marketplace components.
Scalability: Workflows should handle large volumes of files efficiently.
Robustness: Strong error-handling and logging mechanisms.
Documentation: Clear and user-friendly workflow guides.

Submission Deadline
Task Deadline
Proposal Due 17.12.24
Project Start 06.01.25

Deliverables
UiPath Workflows: Functional workflows for each file type (PDF, Excel, Word, and Email).
Documentation: Clear user guides, workflow diagrams, and troubleshooting notes.
Test Cases: Validation results using sample files provided.

Ongoing Development Opportunities
If successful in this project, the freelancer may be considered for:
Expanding workflows to additional file types (e.g., CAD, GIS).
Enhancing workflows for advanced error handling, compliance mechanisms, and scalability.
Integrating anonymisation processes to ensure GDPR compliance.

By incorporating an hourly rate, this approach ensures flexibility while allowing us to evaluate proposals based on time efficiency, experience, and cost. We look forward to receiving your proposals!

This version emphasizes the hourly rate approach and invites freelancers to provide detailed estimates and methods for each milestone while allowing flexibility for negotiation.

Skills: Machine Learning (ML), Artificial Intelligence, Data Scraping, Data Extraction, Deep Learning
Hourly rate: 25 - 50 USD
3 hours ago
  • Websites, IT & Software, Data Entry & Admin, Engineering & Science, Artificial Intelligence, Data Scraping, Data Extraction, Machine Learning (ML), Deep Learning
Data Engineer
21,000 USD 3 hours ago
Client Rank - Medium

Payment method verified
40 jobs posted
4 open job
no reviews
Registered at: 09/05/2024
IN India
Medium
Required Connects: 10
Senior Data Engineer (Snowflake/ETL)
Experience: 5+ Years
What makes you a great fit:
 You have 5+ years of extensive development experience using snowflake or
similar data warehouse technology
 You have working experience with dbt and other technologies of the modern
datastack, such as Snowflake, Apache Airflow, Fivetran, Looker, AWS, git.
 You have experience in agile processes, such as SCRUM
 You have extensive experience in writing advanced SQL statements and
performance tuning them
 You have experience in Data Ingestion techniques using custom or SAAS tool
like fivetran
 You have experience in data modelling and can optimise existing/new data
models
 You have experience in data mining, data warehouse solutions, and ETL, and
using databases in a business environment with large-scale, complex datasets
 You having experience architecting analytical databases (in Data Mesh
architecture) is added advantage
 You have experience working in agile cross-functional delivery team
 You have high development standards, especially for code quality, code reviews,
unit testing, continuous integration and deployment
As a Analytics Engineer you’ll be:
 Developing end to end ETL/ELT Pipeline working with Data Analysts of business
Function.
 Designing, developing, and implementing scalable, automated processes for
data extraction, processing, and analysis in a Data Mesh architecture
 Mentoring other Junior Engineers in the Team
 Be a “go-to” expert for data technologies and solutions
 Ability to provide on the ground troubleshooting and diagnosis to architecture
and design challenges
 Troubleshooting and resolving technical issues as they arise
 Looking for ways of improving both what and how data pipelines are delivered
by the department
 Translating business requirements into technical requirements, such as entities
that need to be modelled, DBT models that need to be build, timings, tests and
reports
 Owning the delivery of data models and reports end to end
 Perform exploratory data analysis in order to identify data quality issues early in
the process and implement tests to ensure prevent them in the future
 Working with Data Analysts to ensure that all data feeds are optimised and
available at the required times. This can include Change Capture, Change Data
Control and other “delta loading” approaches
 Discovering, transforming, testing, deploying and documenting data sources
 Applying, help defining, and championing data warehouse governance: data
quality, testing, coding best practices, and peer review
 Building Looker Dashboard for use cases if required
Skills: Fivetran, Looker, AWS Application, Git, SaaS, ETL
Fixed budget: 21,000 USD
3 hours ago
  • Data Science & Analytics, Data Mining & Management
Data Engineer(Snowflake & Looker)
21,000 USD 3 hours ago
Client Rank - Medium

Payment method verified
40 jobs posted
4 open job
no reviews
Registered at: 09/05/2024
IN India
Medium
Required Connects: 10
Senior Data Engineer (Snowflake/ETL)
Experience: 8+ Years
What makes you a great fit:
• You have 5+ years of extensive development experience using snowflake or
similar data warehouse technology
• You have working experience with dbt and other technologies of the modern
datastack, such as Snowflake, Apache Airflow, Fivetran, Looker, AWS, git.
• You have experience in agile processes, such as SCRUM
• You have extensive experience in writing advanced SQL statements and
performance tuning them
• You have experience in Data Ingestion techniques using custom or SAAS tool
like fivetran
• You have experience in data modelling and can optimise existing/new data
models
• You have experience in data mining, data warehouse solutions, and ETL, and
using databases in a business environment with large-scale, complex datasets
• You having experience architecting analytical databases (in Data Mesh
architecture) is added advantage
• You have experience working in agile cross-functional delivery team
• You have high development standards, especially for code quality, code reviews,
unit testing, continuous integration and deployment
As a Analytics Engineer you’ll be:
• Developing end to end ETL/ELT Pipeline working with Data Analysts of business
Function.
• Designing, developing, and implementing scalable, automated processes for
data extraction, processing, and analysis in a Data Mesh architecture
• Mentoring other Junior Engineers in the Team
• Be a “go-to” expert for data technologies and solutions
• Ability to provide on the ground troubleshooting and diagnosis to architecture
and design challenges
• Troubleshooting and resolving technical issues as they arise
• Looking for ways of improving both what and how data pipelines are delivered
by the department
• Translating business requirements into technical requirements, such as entities
that need to be modelled, DBT models that need to be build, timings, tests and
reports
• Owning the delivery of data models and reports end to end
• Perform exploratory data analysis in order to identify data quality issues early in
the process and implement tests to ensure prevent them in the future
• Working with Data Analysts to ensure that all data feeds are optimised and
available at the required times. This can include Change Capture, Change Data
Control and other “delta loading” approaches
• Discovering, transforming, testing, deploying and documenting data sources
• Applying, help defining, and championing data warehouse governance: data
quality, testing, coding best practises, and peer review
• Building Looker Dashboard for use cases if required
Skills: ETL, Looker, Fivetran, AWS CloudFormation, BigQuery, Tableau, Big Data, Data Analysis
Fixed budget: 21,000 USD
3 hours ago
  • Data Science & Analytics, Data Mining & Management
Web Scraping Specialist Needed for Data Extraction Project
30 - 70 USD / hr
3 hours ago
Client Rank - Risky

Payment method not verified
1 open job
no reviews
Registered at: 09/12/2024
MK Macedonia
Risky
Required Connects: 9
We are seeking a skilled and experienced Web Scraping Specialist to join our team on a freelance/contract basis. The ideal candidate will have a strong background in web scraping, data extraction, and data parsing, as well as a track record of delivering accurate and efficient results.

Responsibilities:

Develop and deploy web scraping scripts to extract data from various websites and APIs.
Parse, clean, and format the scraped data into usable formats (e.g., CSV, JSON, Excel).
Ensure compliance with website terms of service and legal guidelines.
Optimize scripts for efficiency and handle large-scale data extraction.
Troubleshoot and maintain scraping tools to ensure reliability.
Collaborate with our team to understand project requirements and deliverables.
Requirements:

Proven experience in web scraping using tools like Python (BeautifulSoup, Scrapy, Selenium), Puppeteer, or similar technologies.
Strong understanding of HTML, CSS, JavaScript, and DOM manipulation.
Ability to work with APIs and manage authentication mechanisms.
Familiarity with data storage solutions and database integration.
Attention to detail and commitment to delivering high-quality results.
Knowledge of ethical scraping practices and legal compliance.
Strong problem-solving skills and ability to work independently.
Preferred Qualifications:

Experience with cloud platforms for scaling web scraping tasks (e.g., AWS, Google Cloud, Azure).
Familiarity with proxy rotation and anti-bot bypass techniques.
Background in data analysis or visualization is a plus.
Project Scope:

[Describe the specific data source(s) and the type of data you need.]
[Indicate the volume of data and frequency of extraction.]
[Mention any specific format or integration required.]
What We Offer:

Competitive compensation based on project scope and complexity.
Flexible deadlines and collaborative working environment.
Opportunity for ongoing projects and long-term collaboration.
If you’re a detail-oriented problem solver with a passion for data, we’d love to hear from you! Please send us your portfolio, relevant experience, and any sample work showcasing your web scraping expertise.
Skills: Data Scraping, Data Mining, Python, Data Extraction, Scrapy
Hourly rate: 30 - 70 USD
3 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Data Extraction & Analyticis for underwriting reports
not specified 3 hours ago
Client Rank - Medium

Payment method verified
$150 total spent
1 hires
1 jobs posted
100% hire rate, 1 open job
5.00 of 1 reviews
Registered at: 13/01/2020
US United States
Medium
Required Connects: 11
Develop data sets and high-level reports for underwriting loss ratio & underwriting characteristics in an aggregate form. Looking for the ability to break out such reports to show loss ratio by insurance carrier, by coverage form, by location, etc.
Skills: ETL, Query Tuning, C#, Oracle PLSQL, Visual Basic for Applications, Microsoft SQL Server Programming, SQL Programming, SQL Server Integration Services, Microsoft Excel, Database Modeling, Microsoft Project, Analytics, Microsoft Access Programming, Data Warehousing
Budget: not specified
3 hours ago
  • Web, Mobile & Software Dev, Other - Software Development
Data Mining & List Scraping Specialist
5 USD 3 hours ago
Client Rank - Excellent

Payment method verified
$8'051 total spent
293 hires , 12 active
306 jobs posted
96% hire rate, 58 open job
6.23 /hr avg hourly rate paid
196 hours
5.00 of 378 reviews
Registered at: 22/03/2024
GB United Kingdom
Excellent
Required Connects: 10
We are seeking a skilled Data Mining and List Scraping Specialist to compile a detailed database of music agencies in Switzerland. Your responsibilities will include researching and collecting the following information:

Agency Name
Street Address
Zip Code & City
Email Address
Phone Number
Website URL
Main Language
Business Manager
Social Media Links (Facebook, Instagram)
Skills: Data Scraping, Lead Generation, List Building, Data Mining, Data Entry, Microsoft Excel, Data Extraction
Fixed budget: 5 USD
3 hours ago
  • Data Science & Analytics, Data Mining & Management
Brevo Contacts Export - Brevo API Export
30 USD 4 hours ago
Client Rank - Excellent

Payment method verified
$77'046 total spent
116 hires , 33 active
157 jobs posted
74% hire rate, 2 open job
16.88 /hr avg hourly rate paid
2517 hours
4.94 of 83 reviews
Registered at: 15/03/2013
US United States
Excellent
Required Connects: 10
We're aiming to switch from Brevo to Active Campaign. To do this, we need to export our Brevo contacts to our new Active Campaign account.

We need the exports to be organised in our existing lists.

Brevo doesn't allow you to export more than 100 contacts from any list at one time, without using their API:

https://developers.brevo.com/reference/requestcontactexport-1

We need someone who is able to export all of our lists at once using any time-efficient method.

IMPORTANT TO NOTE:

We only want to export contacts that have opened an email in the last 12 months, and we want all their attributes exported.
Skills: Data Scraping, API, Data Mining, Data Extraction, API Integration
Fixed budget: 30 USD
4 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Data Scraping into Spreadsheet
not specified 4 hours ago
Client Rank - Excellent

Payment method verified
$13'440 total spent
10 hires , 7 active
9 jobs posted
100% hire rate, 1 open job
8.70 /hr avg hourly rate paid
1404 hours
5.00 of 3 reviews
Registered at: 29/08/2023
GB United Kingdom
Excellent
Required Connects: 13
We will provide you a spreadsheet and a list of websites that we want you to take the data from and enter into the spreadsheet.
Skills: Data Scraping, Data Entry, Data Mining, Data Extraction
Budget: not specified
4 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Website Scraper Needed for Data Extraction
3 - 7 USD / hr
5 hours ago
Client Rank - Excellent

Payment method verified
$280'556 total spent
171 hires , 43 active
336 jobs posted
51% hire rate, 1 open job
8.74 /hr avg hourly rate paid
27602 hours
4.78 of 111 reviews
Registered at: 30/12/2011
GB United Kingdom
Excellent
Required Connects: 19
We are seeking an experienced website scraper to help us extract specific information from a specific website and compile it into a structured spreadsheet. The ideal candidate should have a strong background in web scraping techniques and be proficient in tools such as Python, Beautiful Soup, Scrapy, or whatever tool can do the job with the minimum of fuss. Your role will involve identifying the necessary data points, developing scripts, and ensuring the accuracy of the extracted data. If you have a keen eye for detail and can deliver results efficiently, we would love to hear from you!

You will be asked to provide a sample extraction of 10 records to prove you can do the job and to estimate total time required.

Please ensure you mention the word IMROI as the first word in your response otherwise your application will be rejected.
Skills: Data Scraping, Data Extraction, Scrapy, Data Mining
Hourly rate: 3 - 7 USD
5 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Create brand imagery for new AI Web Scraping focused landing page
100 USD 5 hours ago
Client Rank - Excellent

Payment method verified
$59'188 total spent
19 hires , 9 active
14 jobs posted
100% hire rate, 1 open job
37.99 /hr avg hourly rate paid
269 hours
5.00 of 9 reviews
Registered at: 29/04/2020
FR France
Excellent
Required Connects: 16
Please create a set of brand images (5 images) that will sit on a landing page which focuses on our AI Web Scraping feature. The main image will sit above the fold.

The different images should highlight the different features of the AI web scraping functionality.

Here's the documentation for the AI web scraping feature:
https://www.scrapingbee.com/documentation/#data-extraction-with-ai-beta

It should be in the same style as our existing brand imagery which I've attached.

The finished images should be in svg format.
Skills: Website Asset, Adobe Illustrator, Adobe Photoshop, Brand Identity & Guidelines, Graphic Design
Fixed budget: 100 USD
5 hours ago
  • Design & Creative, Graphic, Editorial & Presentation Design
Data Scraping from Health Suppliers Website
5 - 10 USD / hr
5 hours ago
Client Rank - Medium

Payment method verified
1 open job
no reviews
Registered at: 05/11/2024
NL Netherlands
Medium
Required Connects: 9
We are seeking a skilled data scraper to extract information from a health suppliers website. The website contains a comprehensive list of health suppliers, including GPs, hospitals, pharmacies, and more. The ideal candidate should have experience in web scraping and be proficient in handling large datasets. Your task will involve gathering relevant data points and organizing them into a structured format for further analysis. Please provide examples of your previous web scraping projects when applying.

Scraping information:
- Company name
- Owners name
- Location (city and province)
- phone number
- E-mail
- Website

Scraping quantities:
- 29.201 companies
- 1462 pages
- 1 website

**Relevant Skills:**
- Web scraping
- Data extraction
- Data cleaning
- Proficiency in Python or similar languages
- Experience with libraries like BeautifulSoup or Scrapy
Skills: Data Scraping, Data Entry, Data Mining
Hourly rate: 5 - 10 USD
5 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Public Database Scraper (API?)
not specified 5 hours ago
Client Rank - Medium

Payment method verified
7 jobs posted
2 open job
no reviews
Registered at: 13/02/2023
GB United Kingdom
Medium
Required Connects: 10
Our brand wants to provide Free Reports to users who sign up on our landing page. They will provide their US zipcode.

The Reports will be very brief, and summarised in 5-10 bullet points in an email.

All the data required from the reports are available on the Environmental Working Group (EWG) database, which is also specific to zipcodes.

I'm aware that existing brand are doing this, I suspect through an API of sorts.

If you think this is a problem you can solve, let's talk.
Skills: API, Data Scraping, Data Extraction, Data Mining
Budget: not specified
5 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Listex scrape
15 USD 6 hours ago
Client Rank - Medium

Payment method verified
$931 total spent
10 hires , 4 active
8 jobs posted
100% hire rate, 4 open job
5.00 of 4 reviews
Registered at: 25/07/2022
LT Lithuania
Medium
Required Connects: 10
Hi there, look for a dev to scrape this website:

https://listex.info/uk/food-beverage/ (Only food category)

Data needed in original languages, exported to an excel file:

- Product title (if available find it under "Альтернативне найменування (англ)", if not then any name that has this value "англ" near it. If that's also missing, any language avaliable. )

- Product title language (The language the product title was extracted. If it was (англ), then (англ) needs to be in the cell.

- Product barcode (under Вагогабаритні характеристики, the very first barcode)

- Product URL

- Ingredient list ("Склад (оригінал)". If original is not avaliable, etract the ingredient list under "Склад (укр.)" or
"Склад (рос.)"

- Ingredient list language (The language the ingredient list was extracted. If it was (англ), then (англ) needs to be in the cell.

- Allergen list (Text is under "Алергени". Only extract allergen names that have "ТАК", or Yes near the type of allergen)

- Category

- Product image URL

- Certifications that product is vegan, vegetarian, gluten free, etc (Find it under Основні. Only extract values that have "ТАК", or Yes near the certification)

Also need to receive the script once done
Skills: Data Scraping, Data Extraction, Data Mining
Fixed budget: 15 USD
6 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Data Extraction from Air Quality Platforms
6 - 16 USD / hr
6 hours ago
Client Rank - Risky

Payment method not verified
1 open job
PH Philippines
Risky
Required Connects: 7
I need a freelancer to help access and extract historical air quality data from platforms like Plume Labs or similar APIs. Tasks include:
- Setting up API integration for data collection.
- Automating the process (e.g., Python scripts, Google Sheets).
- Delivering the data in a clean format (CSV/Excel).
Skills: Data Extraction, Data Mining, Data Collection
Hourly rate: 6 - 16 USD
6 hours ago
  • Data Science & Analytics, Data Extraction/ETL
Scraping website into Google Sheet
5 USD 6 hours ago
Client Rank - Excellent

Payment method verified
$2'364 total spent
22 hires , 8 active
32 jobs posted
69% hire rate, 2 open job
13.00 /hr avg hourly rate paid
5 hours
5.00 of 13 reviews
Registered at: 15/12/2021
FR France
Excellent
Required Connects: 10
SCRAP AND EXTRACT DATA PROFILES FROM WEBSITE TO GOOGLE SHEET

The objective is to scrape and extract data from a job board. Below are the key steps and expectations from this process:
**WATCH VIDEO**
https://www.loom.com/share/465c33be3d1b4dd9847c1fd6310e72e6?sid=f5eb7657-96d7-4d50-b9fc-3273ccd8a09c

Very important: please watch this video first
**Process Overview ** :

1. Access the Website: Use the specific access I have created for you to log into the website, which is in French. No translation is needed.
2.Data Extraction: Navigate to the specified page where you can use filters to sort the data (Yes/No). Begin with the 'Yes' filter, where approximately 8,000 entries are listed, and then proceed with the 'No' filter.
3.Scraping Details: For each profile, capture the following details: (see Google Sheet)
4.Handling Incomplete Data: If certain information is not available on a profile, label it as 'blank' or 'not available'.

Template and Instructions: I will share a template with you to standardize the data entry process. Ensure all the required fields are filled as per the instructions in the video.

**Special Notes**:
• No need to Download CVs. Just need the Links of the CVs as the website hosts these documents.
• Focus on extracting detailed profile information as outlined, including any available presentations or summaries.

Before proceeding, please answer the following four questions:
1. What tools would you recommend for this task?
2. How quickly can you set up the entire automation process?
3. How long will it take to finalize the entire scraping and data extraction process?
4. How much do you charge for the full service of setting up and extracting the data?

Your responses to these questions are crucial for us to proceed with your services.
Skills: Data Scraping, Data Entry, Google Sheets, Google Docs, Data Mining, Data Extraction, JavaScript, Scrapy
Fixed budget: 5 USD
6 hours ago
  • Data Science & Analytics, Data Extraction/ETL
PDF Image to Text & Data Extraction
~29 USD / hr
6 hours ago
Client Rank - Risky

Payment method not verified
1 open job
no reviews
Registered at: 10/12/2024
IN India
Risky
I'm looking for an expert to extract both text and images from PDF files. The extracted data should be organized into a plain text document.

Key Requirements:
- Proficient in data extraction from PDF files
- Able to handle both text and images
- Capable of organizing data into a plain text document

The primary use of the extracted data will be for analysis and processing. Therefore, attention to detail and accuracy in data extraction is crucial.

Please extract images as individual files. The extracted images should be saved as JPEG files. The total number of PDF files or pages to be extracted is between 10 and 50. Please name the extracted images using the PDF name followed by the page number (e.g., document1_page1, document1_page2).

Please organize the extracted text data by document. Please ensure basic level detail extraction. Please complete the project within 1 week. You can use any tool of your preference for the data extraction. Please output the text data as a plain TXT file. You can use any suitable modern software for data extraction.

Skills: Data Processing, Data Entry, Excel, Data Mining, Copy Typing
Hourly rate: 2500 INR
6 hours ago
  • Writing & Content, Data Entry & Admin, Engineering & Science, Copy Typing, Data Processing, Data Entry, Excel, Data Mining
PDF Image to Text & Data Extraction
~29 USD / hr
6 hours ago
Client Rank - Risky

Payment method not verified
1 open job
no reviews
Registered at: 10/12/2024
IN India
Risky
I'm looking for an expert to extract both text and images from PDF files. The extracted data should be organized into a plain text document.

Key Requirements:
- Proficient in data extraction from PDF files
- Able to handle both text and images
- Capable of organizing data into a plain text document

The primary use of the extracted data will be for analysis and processing. Therefore, attention to detail and accuracy in data extraction is crucial.

Please extract images as individual files. The extracted images should be saved as JPEG files. The total number of PDF files or pages to be extracted is between 10 and 50. Please name the extracted images using the PDF name followed by the page number (e.g., document1_page1, document1_page2).

Please organize the extracted text data by document. Please ensure basic level detail extraction. Please complete the project within 1 week. You can use any tool of your preference for the data extraction. Please output the text data as a plain TXT file. You can use any suitable modern software for data extraction.

Skills: Data Processing, Data Entry, Excel, Data Mining, Copy Typing
Hourly rate: 2500 INR
6 hours ago
  • Writing & Content, Data Entry & Admin, Engineering & Science, Copy Typing, Data Processing, Data Entry, Excel, Data Mining
Call to action
Freelancing is a business
Make it more profitable with Vollna

Streamline your Upwork workflow and boost your earnings with our smart job search and filtering tools. Find better clients and land more contracts.