
Scraping Projects
Looking for freelance Scraping jobs and project work? PeoplePerHour has you covered.
Data scraping
I require an experienced data researcher / scraping specialist to extract high-quality, niche-specific data from multiple UK sources and deliver it in a clean, structured format. This is not a bulk scraping job — accuracy and relevance are critical. Scope of Work: 1. Target Data (Initial Focus) I am looking to build lists for the following niches: • Property investors (Buy-to-Let, HMO, developers) • Company directors in property-related businesses • Business owners suitable for commercial finance ⸻ 2. Data Sources You may use a combination of: • Companies House (via SIC codes & director data) • Property-related platforms (where legally accessible) • Public directories • LinkedIn (Sales Navigator filtering – manual research, not scraping if restricted) • Other compliant UK data sources ⸻ 3. Required Data Fields Each record should include (where available): • Full Name • Company Name • Role (Director / Owner) • Email Address (verified where possible) • Phone Number (if available) • Location (UK-based) • Industry / SIC Code • Notes (if relevant, e.g. property investor / developer) ⸻ 4. Data Quality Requirements • UK-based contacts only • No duplicate entries • No generic or irrelevant businesses • Data must be accurate and up-to-date • Avoid scraped junk or low-quality lists ⸻ 5. Compliance (VERY IMPORTANT) • All data must be sourced from publicly available or compliant sources • Must comply with UK GDPR guidelines • No illegal scraping or data extraction methods • No use of prohibited LinkedIn automation tools ⸻ 6. Output Format • Excel / Google Sheets • Clearly structured columns • Clean, ready for outreach use ⸻ 7. Initial Volume • Phase 1: 4000-5000 high-quality leads • Potential for ongoing weekly/monthly work ⸻ Ideal Candidate: • Experience with UK data sourcing (Companies House, etc.) • Strong data cleaning and validation skills • Familiar with B2B lead generation • Understanding of property / finance sector is a plus • Able to suggest better data sources (not just follow instructions) ⸻ Budget & Timeline: • Open to proposals based on quality • Looking to start immediately • Ongoing work available for the right person ⸻ To apply, please include: • Examples of similar data projects • Tools you use for sourcing and verification • Your approach to ensuring data accuracy ⸻ Start your proposal with: “Quality over quantity — understood” ⸻ Additional Notes: This project is focused on building a long-term data pipeline for targeted outreach campaigns. I am looking for someone reliable who can consistently deliver high-quality results.
4 days ago44 proposalsRemoteData scraping for influencers (automation)
Hello I’m looking for data for a specific set of influencers Specifically for influencers with 75K+ followers on Instagram or 50k subs on YouTube With an audience over 40 The influencer will be speaking on topics like financial planning, many wellbeing categories. Pls tell me how you would do this and how much you’ll charge for the data Also can we verify the engagement rate? Say per 1,000 emails and verify the emails? A Python script using the YouTube Data API v3 (free, official) and Phantombuster (Instagram) to automatically search, filter, and export influencer data to a Google Sheet or CSV. YouTube (API — Free & Clean) Use: YouTube Data API v3 Script logic: 1. Loop through keyword list 2. Search type=channel for each keyword 3. Pull channel details: name, URL, subscriber count, description, email (regex parse from description) 4. Filter: subscribers between 50,000 and 500,000 5. Export all results to CSV Keywords to loop: retirement planning lifestyle retirement vlog over 60 grandparenting tips grandma empty nest life over 50 pickleball over 50 senior fitness over 60 menopause health women over 50 midlife women wellness caregiver aging parents RV retirement full time gardening over 50 vegetable men's health over 50 medicare senior health Christian women over 50 faith retirement lifestyle active aging senior Output columns: Channel Name | URL | Subscribers | Email | Description snippet | Keyword searched Instagram (Phantombuster) Use: Phantombuster Instagram Hashtag Scraper + Profile Scraper (client will provide API key and session cookie) Script/phantom logic: 1. Run Instagram Hashtag Post Scraper on each hashtag below 2. Pull 500 posts per hashtag → extract unique profile URLs 3. Feed profile URLs into Instagram Profile Scraper 4. Extract: username, follower count, bio text, email from bio, website URL 5. Filter: followers 50,000–500,000 6. Export to CSV Hashtags to scrape: #retirementplanning #retirementlife #retireearly #grandparents #grandmalife #nanalife #grandpalife #emptynesters #lifeafterkids #over50life #pickleballlife #activeaging #seniorfitness #womenover50 #menopause #midlifewomen #caregiverlife #agingparents #seniorcare #rvlife #fulltimerv #retirementtravel #vegetablegarden #gardeningover50 #growyourown #menshealth #healthyaging #over50fitness #medicare #seniorliving #agingwell #christianwomen #faithoverfear #christianliving Output Format Single CSV, one row per creator: Platform | Handle/Channel | URL | Followers | Email | Bio/Description | Category | Source Keyword Filtering Rules (build into script) ∙ Followers: 50,000–500,000 only ∙ Skip accounts with 0 posts or last post over 60 days ∙ Skip accounts where bio contains: “under 18”, “teen”, “student”, “college” ∙ Flag rows with no email (for Apollo enrichment pass) ∙ Deduplicate on URL before export Deliverable
7 days ago18 proposalsRemotePython script run and scrape data
You need to build the script and run the data I don’t mind the process but the outcome is the data. Most likely using these methods Develop a Python automation to harvest influencer data from YouTube (Data API v3) and Instagram (Phantombuster) and export a deduplicated CSV. For YouTube: iterate provided keywords, search channels, capture channel name, URL, subscriber count, description snippet, parse emails, filter 50k–500k subscribers, and log keyword source. For Instagram: run hashtag scrapers, collect up to 500 posts per tag, extract unique profiles, scrape username, followers, bio, email, website, apply follower and activity/bio filters, flag missing emails, and produce a unified CSV with specified columns and filtering rules.
7 days ago25 proposalsRemoteI need a magneto e-commerce site scrapping for products
Our supplier uses magneto and we wish to scrape all the product data including images to upload onto our own site. Due to the sheer volume of products this is the easiest way to achieve this. The products then need uploading to our own woocommerce site with the image titles changing to remove the suppliers name and made SEO friendly.
a day ago27 proposalsRemoteEmail addresses for UK Companies in specific industries
Good afternoon, I am looking for a freelancer who can help with sourcing data files of email addresses for actively trading companies in England who are involved in the wholesale meat supply trade. I'm not looking to buy a random data list, these companies must be actively trading and fall under one of the following SIC codes: 46320, 10110, 10120, 10130, 47220. If you don't know what SIC codes are, this job is not for you. I am a freelance professional myself but this falls out my area of expertise. If this is a job you think you can reliably deliver, there is great potential for a long-term relationship. Along with your proposal, please outline how you plan to deliver this project, are you running a script, scraping web pages, buying a list? Also please provide an example of 10-20 data samples so I know that you understand what I am looking for. I haven't set a budget for this, but am willing to negotiate with a reliable professional for clean data. Budget is a placeholder only.
3 days ago36 proposalsRemoteUK Business DATA Supplier -
I am looking for a business data supplier. Data will be independant businesses - owners name, business name, address, email, whatsapp , post code price per 1000, 10000 & 100000 + turn around time. If you can scrape any other information for direct marketing, please let us know, including LinkedIn & plastic card companies Regards Proactiv
25 days ago27 proposalsRemoteProspect Intelligence Analyst | Research Assistant
ROLE OVERVIEW Our firm helps small service businesses in the US and UK identify and fix operational revenue leaks — the gaps that cause them to lose enquiries and bookings without realizing it. You sit at the front of our Prospect Machine. Each week you research small businesses, identify their primary revenue leak, score them, enrich decision maker contacts, and populate our structured prospecting tracker. Your output feeds directly to our Cold Caller and Business Systems Consultants. This is not a data entry role. It requires pattern recognition, fast decision-making from limited information, and the discipline to work at consistent pace to a fixed weekly deadline. CORE RESPONSIBILITIES • Source 150–200 raw businesses per week using scraping tools provided by company • Filter to 120–150 qualified SMBs • Review each business's online digital presence — website, booking system, social pages, and online reputation — and conduct test calls outside office hours to assess missed call risk and after-hours responsiveness. Identify the primary operational revenue leak based on what the evidence shows • Find and verify the decision maker via enrichment tools provided by company • Score each lead, flag Priority Leads, and escalate leads immediately • Populate the B2B Prospecting Tracker • Submit all deliverables via agreed upon platform and time. Onboarding Ramp Week 1— 50–60 leads, research and tracker only. Full SOP and training provided. Quality standards apply from day one. Week 2 onwards — Cold call script prep for top 20 Priority Leads added. Week 3 onwards — Industry community and directory identification added. REQUIREMENTS • Experience in B2B lead research, business intelligence, or structured data research • Able to make fast, confident decisions from publicly available data • Strong attention to detail — accurate entries matter more than perfect ones • Comfortable following a structured SOP independently, without frequent check-ins • Clear written and spoken English, with prompt communication COMPENSATION & STRUCTURE • $17.00/hr · 15 hrs/week · ~$255/week • Performance review at 45 days — rate increase available for strong performers HOW TO APPLY Begin your application with the word SIGNAL — applications that don't will not be reviewed. Then answer these two questions: 1. Describe a research or data project where you worked to a consistent weekly output target. What tools did you use and what was your weekly volume? 2. You're researching a dental practice. Their website has no online booking system — new patients are instructed to call during office hours only. There is no contact form and no alternative way to enquire outside of calling. In two to three sentences: identify the primary revenue leak, explain why it matters commercially, and give this lead a score out of 10 with a one-sentence justification. Note Question 2 has a clear correct answer. We are looking for specific, evidenced reasoning — not a general description of the problem.
15 days ago12 proposalsRemote
Past "Scraping" Projects
opportunity
Python Data Pipeline — Web Scraping, Multi-Platform, languages
I'm building a children's activity discovery platform for Switzerland (think "Google for kids' activities"). I need a recurring data pipeline that scrapes class schedules from ~500 providers across multiple booking platforms in Geneva, expanding to all of Switzerland within 18 months. What needs to be scraped: Cogito-Sport (Angular/JavaScript portal) — 8-12 swimming clubs iClassPro (JavaScript portal) — 3-5 providers loisirsjeunes.ch (static HTML, paginated, ~200 activities via sequential IDs) Ville de Genève sports index (static HTML) PDF timetables (22 community centres) Individual club websites (mixed HTML) What to extract from each source: Activity name, provider, day of week, time, age range, price, address, registration URL. I'm building a children's activity discovery platform for Switzerland (think "Google for kids' activities"). I need a recurring data pipeline that scrapes class schedules from ~500 providers across multiple booking platforms in Geneva, expanding to all of Switzerland within 18 months. What needs to be scraped: Cogito-Sport (Angular/JavaScript portal) — 8-12 swimming clubs iClassPro (JavaScript portal) — 3-5 providers loisirsjeunes.ch (static HTML, paginated, ~200 activities via sequential IDs) Ville de Genève sports index (static HTML) PDF timetables (22 community centres) Individual club websites (mixed HTML) What to extract from each source: Activity name, provider, day of week, time, age range, price, address, registration URL. Deliverables: Working scrapers for all source types Config file to add new providers without code changes Normalisation layer mapping all sources to unified schema Change detection and summary email Deployment on Railway with cron schedule Clean CSV output + optional Airtable API push README written for non-developers In your proposal, briefly describe how you have handled JavaScript-rendered Angular or React pages in a previous project — what tools did you use and how did you handle DOM waiting? I have a full technical spec document available on request. P.S. Please note: due to payment processing limitations I am unable to work with freelancers based in Russia or Belarus.
Web Scraping Required
I require a structured data extraction project from the following directory: https://www.buildington.co.uk/companies The objective is to extract and structure company data into a clean Excel spreadsheet. Required fields: • Company Name • Contact Name (if available) • Telephone Number • Email Address • Website URL Important: • Some data is available directly on the directory page. • In certain cases, the freelancer may need to visit the company’s website to retrieve missing contact details. • Data must be structured, cleaned and deduplicated. • Output format: Excel (.xlsx) with clearly labelled columns. • Please confirm your approach and tools before starting. This is not a one-off copy and paste task. I am looking for someone who can create a reliable and efficient extraction method.
Website Product Availability Scanner
Simple search to be fetched & send it to an email once every day at fixed time Seeking a freelancer to develop a program that scans a website daily for a specific product and sends an email notification regarding its availability. The program should be efficient and reliable, ensuring timely updates. Deliverables Develop a web scraping program Implement email notification system Ensure daily execution of the program More accurate details will be given to freelancer as it will hardly be 1-3 hours of work
Data Scraping
We are looking for a freelancer with experience in data extraction and web automation to collect a list of registered businesses from a Laravel-based platform that requires login. I have valid login credentials (my own account). The task includes: Logging in using provided credentials Accessing the authenticated business listing Handling pagination to retrieve all entries Exporting the data to CSV or Excel
opportunity
Database scraped of UK limited companies 2
Hi there We need a database scraped of UK limited company names, their UK mobile number beginning 07, their email address, their website address, the Facebook business page URL and the link from where the details were obtained. We require any amount up to 5,000,000 records. Thanks
opportunity
Facebook Groups Scraping
Hi there We are looking for a freelancer to scrape the details of people who are members of open groups on Facebook and are business owners, and then collect the information of their business via their personal Facebook profile links to their business page. Thanks
Attendee list data collection into a csv
Hi, I need a web page scraping for an event attendee's details. There will be approx 15k records in total so 15 rows Columns to be: First name Last name Job Title Company Country Email address Phone number About me Strict deadline of Tuesday 10th 5pm UK time. I will provide the log in to the portal
opportunity
Product Scrape and Data Extration
I need the following 4 categories scraped and data extracted. https://www.partstown.com/wmf/parts - 768 Products https://www.partstown.com/franke/coffee-maker-parts/parts - 85 Products https://www.partstown.com/franke-foodservice-coffee/coffee-maker-parts/parts - 53 Products https://www.partstown.com/franke-foodservice-system/coffee-maker-parts/parts - 100 Products I need the following items in a Google Sheet or Excel sheet: SKU (Parts Town #:) Title List Price Models Fits (pipe separated if multiple) Product Description 1st Image URL Stock Status
Data scraping needed
I have several directories I need to obtain contact information from... Can anyone help me? I will have all the sites for the person who will help!
Single airline Flight Data web scrape
I need someone to provide me with a list of all Emirates flights on 21st November 2025. The list must include Point of departure, destination, flight time, aircraft type and callsign. AI tells me that Emirates operates around 500 individual flights per day. The data can be obtained from Flight Radar24 for which I can provide a temporary Gold subscription log in. I have provided a blank copy of the excel spreradsheet that needs completing with the data.
opportunity
Time-Sensitive Web Form Automation
We require an experienced automation engineer to build a high-precision browser automation workflow for submitting a time-sensitive web application form. The form opens daily at a fixed time and has limited availability. The objective is to optimise submission timing and reduce manual latency. Scope Pre-load and securely store required form data Launch browser session in advance Synchronise system clock via NTP Trigger submission at an exact timestamp Handle potential page load delays or retries Log response times and outcomes Technical Requirements Understanding of browser event timing Clean, maintainable code Respectful request handling (no scraping / no excessive request flooding) Important The solution must operate within normal browser behaviour and must not attempt to bypass authentication systems, rate limiting, or security protections. Source code to be provided on delivery.
opportunity
Elite List Building
Senior Market Researcher Needed: Elite List Building (Deep Verification & Psychographic Profiling Required) We are looking for a meticulous Senior Market Researcher to build a high-value "Market Entry Dossier" of key decision-makers. ⚠️ READ BEFORE APPLYING: This is not a bulk scraping job. If you rely solely on automated tools like Apollo/ZoomInfo without manual verification, do not apply. We require 100% human verification and analysis for every single entry. The Scope: You will be building a targeted list of decision-makers, but your value lies in the quality of the data, not the quantity. Mandatory Requirements: "Active Role" Check: You must verify the contact is currently in the role via LinkedIn activity (posts/comments in the last 90 days). Old data is rejected. Strict Email Validation: No "catch-all" or "risky" emails. You must use SMTP verification tools (NeverBounce/ZeroBounce) for every contact. Psychographic Profiling: You must analyze the prospect’s recent LinkedIn activity to categorize their personality type (Driver, Influencer, Stabilizer, Analyst) and write a specific note on how to approach them. The "First 5" Protocol: You will be required to submit the first 5 rows for approval before continuing. If these do not meet the SOP, the contract will be cancelled immediately. Deliverables: Clean CSV format (No merged cells/formatting). Columns must include: Verified Email, LinkedIn URL, Activity Status, and Personality Insight Note. To Apply: Please start your proposal with the word "DOSSIER" so I know you read the requirements. Tell me which SMTP verification tool you use and one example of how you determine a prospect's personality type from their profile.
opportunity
Extend Existing Scraper to Extract Emails from UK Property Sites
I’m looking for a developer to extend an existing web scraper to work with UK property listing websites in the vacation rental and group accommodation space. The current scraper already: visits websites follows contact pages extracts emails What’s needed is to: 1. Crawl property listing pages 2. Collect property website URLs 3. Reuse the existing email extraction logic on those sites This is an extension of an existing codebase, not a new scraper from scratch. --- Required Skills Python web scraping experience Ability to work with an existing codebase Browser automation for JS-heavy pages Clean, repeatable scraping logic --- Data to Extract Property name or listing title Website URL Email address Phone number (optional) Location (optional) --- Target Niches Holiday lets Vacation houses Group accommodation Large holiday homes --- Expected Notes Not all properties will list emails Some sites use contact forms Output varies by website --- Deliverables Extended scraper using existing code Instructions to run it Clean output (Google Sheets or CSV) --- Notes Repository will be shared after scope is agreed No form submissions No private or gated data