
Data Science Projects
Looking for freelance data science jobs and project work? Browse active opportunities on PeoplePerHour, or hire data scientists through Toptal’s rigorously vetted talent network.
opportunity
API data reporting- Ecommerce marketplaces
Project Title Margin Reporting System – Multi-Source API Data Integration into On-Prem SQL Database Overview We want to build an automated margin reporting system that pulls commercial and operational data from multiple sources via API, standardises it, and stores it in a SQL database hosted on our on-site server. This database will then be used to report margins across products, customers, and orders. The goal is to replace manual reporting and create a reliable “single source of truth” for margin analysis. Objectives Pull data from multiple platforms via API on a schedule Store raw + cleaned data in a structured SQL database Calculate margin metrics consistently Produce reporting-ready tables/views for Power BI / Excel Data Sources (initial) - Shopware (orders, customers, products, pricing) - Exact Globe - Helm / Neuro by despatch company - Ebay / amazon and other marketplaces Key Outputs 1) SQL Database Schema Raw tables (staging layer) Cleaned tables (standardised layer) Reporting layer (margin-ready views/tables) 2) Automated Data Sync Scheduled API pulls (daily, or hourly depending on limits) Incremental updates (not full reload every time) Logging of sync success/failure 3) Margin Logic Margin calculation should support: Revenue (net of VAT) Discounts / vouchers COGS (purchase cost, supplier cost) Shipping income vs shipping cost Payment fees (optional) Returns / refunds / credit notes Deliverables Working data pipeline (API → SQL) Database schema + views Setup instructions for server deployment Documentation of API endpoints used Error handling + logs Handover session (1 hour)
13 days ago54 proposalsRemoteEgocentric Video Data Collection
Project Overview: We are collecting egocentric (first-person POV) video data of everyday household tasks recorded in real residential environments. These videos will be used for human-to-robot / humanoid training datasets. Recording Requirements (Mandatory): First-person (POV) recording using a head-mounted smartphone or equivalent device Recording must reflect natural human viewpoint Handheld, tripod, or surface-mounted recordings are NOT allowed Video Specifications: Resolution: Minimum 1920×1080 (1080p) Orientation: Landscape only Frame Rate: 30 FPS minimum (60 FPS preferred) No slow motion No vertical videos Visibility & Framing Rules: Hands and objects must be clearly visible at all times Camera angle slightly downward No blurred or out-of-focus footage No faces visible (including mirrors, photos, reflections) No IDs, documents, personal data Content Rules: Minimum video duration: 4–5 minutes per video No idle time longer than 3 seconds Only meaningful household tasks (kitchen, living room, bedroom, bathroom.) No eating, drinking, washing hands, sitting idle, or unrelated activities Editing Rules: Raw video files only accepted No trimming No filters No overlays, effects, or text, audio Rejection Criteria - Videos will be rejected if they include: Vertical videos Slow-motion recording Phone held by hand or placed on a surface Camera zoomed in too much Blurry, unfocused, or unstable footage Idle time longer than 3 seconds Edited or trimmed videos Eligibility: Open to international freelancers Must have access to a residential home environment Ability to follow strict technical guidelines is required Before Applying - Applicants must confirm: They have a head-mounted recording setup Their device supports 1080p @ 30 FPS and above. They agree to follow all recording rules strictly Deliverables: Raw video files only Uploaded to assigned cloud folder, sorted by environment Payment Details: Payment only for approved hours and rejected video hours are not paid. Job Types: Part-time, Fresher, Contractual / Temporary, Freelance, Volunteer Contract length: 6 weeks Benefits: Flexible schedule Work from home Work Location: Remote
18 days ago11 proposalsRemoteData Scraping
We are looking for a freelancer with experience in data extraction and web automation to collect a list of registered businesses from a Laravel-based platform that requires login. I have valid login credentials (my own account). The task includes: Logging in using provided credentials Accessing the authenticated business listing Handling pagination to retrieve all entries Exporting the data to CSV or Excel
a month ago33 proposalsRemoteAged payable summary page in Power bi
We provide financial reports to our clients at period end and are currently looking for help with a new report page to display an Aged Payables Summary in Power BI, similar to the format available in Xero. The report needs to be dynamic and respond to the period selected within the Power BI report. The data source is Xero. We use an ETL service to extract the data from Xero and load it into a SQL database, which then serves as the data source for the Power BI report. The report currently utilises the following Xero tables: Journal Invoices Accounts Organisation Tracking Categories we are looking for someone who has already worked on Data from Xero or have created a dynamic Aged payable Summary in Power BI. We would be happy to arrange a meeting to discuss the project requirements in more detail.
6 days ago18 proposalsRemoteopportunity
AI Sales Automation Expert: Build Data & Outreach Workflow
I am looking for an expert in AI-driven sales orchestration to help scale the GTM (Go-To-Market) engine for a specialized industrial product called TorqCut (onshore hydraulic chopsaws for the piling, decommissioning, and demolition markets). What I’ve done so far: I previously built a customized salesflow using: Gemini for deep market and company research. Salesrobot for LinkedIn outreach (I will continue to manage this part myself). Lindy.ai for customized email workflows. The Problem: While the "Deep Research" identified the right companies, the email data was often inaccurate or stale. I am seeing too many "wrong" contacts or bounces, which is wasting automation credits and harming deliverability. The Goal: I need to insert a high-quality service into the middle of this flow that can find and verify high-integrity contact data (specifically Lead Engineers, Procurement Managers, and Operations Directors) before they hit the outreach stage. My Current Thinking: I am currently deciding between Clay and Apollo.io as the data engine. I am leaning toward Clay because of its "waterfall enrichment" and "Claygent" capabilities, which could allow us to scrape prospect websites for specific technical signals (e.g., companies that own 8–24 tonne excavators, which are the required carriers for our TQ-16 to TQ-32 models). I already have Lindy.ai set up and working for the outreach layer, so the solution must integrate seamlessly with it although I am open to better alternative setups to achieve the same result The Ask: I am looking for an expert to: Assess the current stack: Are Apollo or Clay still the best data engines for industrial/construction leads in 2026, or is there a superior AI-first solution I’m missing? Build the "Middle Layer": Set up a robust, automated workflow that takes my research/signals and provides "Golden Records" (95%+ verified emails). Optimize for Scale: I want to reach approximately 1,000 targeted prospects per month at an efficient cost. Integration: Ensure the verified data flows automatically into my existing Lindy.ai sequences or a superior alternative is provided. Important Note: I am fully open to other ideas. If you believe there is a "better way" to achieve this using new autonomous AI SDR agents (like 11x.ai, Artisan, etc.) or a different tech stack that is more efficient or cost-effective than what I’ve proposed, I want to hear your recommendation. Requirements: Proven experience with Clay (waterfall enrichment, Claygent) or Apollo.io. Experience with Lindy.ai or similar autonomous AI agents. Deep understanding of B2B data verification and deliverability (SMTP handshakes, catch-all detection) . Ability to demonstrate how you’ve scaled outreach to 1,000+ leads/mo previously. Please include in your proposal: Which data engine you recommend for the industrial cutting/engineering market and why. A brief overview of how you would structure the workflow from "Signal" to "Sent Email." Estimated monthly tool/software costs for 1,000 verified leads. Looking forward to hearing your ideas!
4 days ago18 proposalsRemoteLinkedin Content Management
I am looking for a sharp, research-capable AI content partner to help me build a credible, insight-driven LinkedIn presence. This is not influencer-style content. This is structured, intelligent, business-focused AI thinking aimed at founders, healthcare executives, investors, and consultants. I need someone who can help with: • Content ideation • Structured research • Drafting short-form LinkedIn posts • Maintaining cadence and quality You are: Strong in AI / technology literacy Comfortable reading technical or policy material Able to think commercially Experienced in LinkedIn content for senior audiences Structured and analytical Able to challenge and elevate ideas Bonus: Experience in healthcare, life sciences, consulting, or B2B strategy Experience ghostwriting for founders or executives
11 days ago39 proposalsRemoteTrailballance and preperation for Audit for small NGO Tanzania
I need someone to llok at our last Audit. Prepare a trial balance and then enter data for the next audit.
6 days ago13 proposalsRemoteData input to a website
I'm a real estate agent. I have about 100 properties that I needed loaded to my website. This will include 100's of photos, property descriptions, property details, etc. I need a reliable freelancer able to do this professionally and accurately. Must be completed quickly
a month ago72 proposalsRemoteI need somone to clean my list of data
I have an email list of approximately 35,000 email addresses and I need someone to clean the data. This can be done either manually or using an automated process — I’m flexible on the method. The requirements are: 1. Remove all duplicate email addresses 2. Remove any invalid email addresses or addresses that are likely to bounce I’m not concerned about how this is carried out, as long as the final list of email addresses is accurate and fully cleaned WITH NO INVALD OR OR EMAILS THAT BOUNCE.
23 days ago72 proposalsRemoteGlobal Project: 300 Native Speakers Needed
We are urgently recruiting native speakers for a large-scale multilingual transcription and AI data project.
5 days ago19 proposalsRemoteUK Part-time Recruitment | WiFi Field Data Collector
UK Part-time Recruitment | WiFi Field Data Collector, Earn Extra Money Easily! Core Highlights • Transparent Salary: 3gbp per asset, more work more gain, no hidden deductions • Flexible Time: No office hours required, take orders and arrange your time freely in spare moments • Ultra-low Threshold: No professional experience needed, easy to master for beginners, with bilingual (Chinese & English) operation manual • UK-wide Coverage: No specific city restrictions, available anywhere in the UK Must-Meet Requirements (All Required) 1. Location Requirement: Must be located in the UK and able to complete on-site data collection tasks locally 2. Device Requirement: • Android smartphone (Android 6.0 or above; Android 9.0 is not recommended as it tends to freeze) • Minimum Memory Requirement: 8+8G (RAM + Storage Memory) • iOS devices are not supported temporarily, please do not apply Key Responsibilities 1. Use the designated APP (WiFi Field Collection APP V1.2) to complete WiFi signal collection at designated locations in accordance with operational standards 2. Complete basic APP operations: take photos, confirm location, submit data (a detailed bilingual operation manual is provided) 3. Maintain stable mobile network and normal positioning during collection to ensure data accuracy Part-time Perks • Salary Guarantee: Transparent and traceable settlement process, no reduction of rights and interests • Easy to Get Started: Step-by-step manual guidance, no skill threshold, accessible to everyone • High Flexibility: No geographical restrictions across the UK, adjust collection locations and time at any time • Standardized Process: Fully transparent task allocation, operational standards and data review Notes 1. Maintain a stable mobile network throughout the work to ensure the APP runs normally and data is synchronized 2. Authorize the APP to access location permissions to assist in accurately locating collection points 3. A complete bilingual (Chinese & English) operation manual will be provided after onboarding, covering the entire process from login to submission Application Method Please send the following information to the designated application channel: [Your Full Name + UK City of Residence + Mobile Phone Model & Android System Version] Review Timeframe: Qualification verification will be completed within 24 hours (focusing on verifying the must-meet requirements) Follow-up Communication: After passing the review, we will inform you of task details, operational standards and salary settlement rules as soon as possible We sincerely invite friends in the UK who meet the device requirements to join us and earn extra money easily in your spare time!
13 days ago1 proposalRemoteopportunity
HEADTEACHERS Primary school email addresses
Please provide email contact data for the following roles within London and Surrey Primary and Secondary Schools: Headteachers Deputy Headteachers Business Managers PE Teachers School Offices Before proceeding, please confirm the total data count and the date of the last data cleanse. It is essential that the data is restricted to London and Surrey only, with no records from other regions. Please also confirm data turnaround time or if data is readily available
20 days ago24 proposalsRemoteopportunity
Macroeconomic Dashboard Design & Development
We are seeking an experienced freelancer to design and develop a comprehensive macroeconomic dashboard to support professional macro and FX analysis. This dashboard will track and present key economic indicators across major economies, including the United States, Australia, New Zealand, Canada, Eurozone, United Kingdom, Japan, and Switzerland. Core indicators will include inflation metrics, GDP and GDP components, central bank interest rates, 2-year government bond yields, and economic data surprise indices. The objective is to create a clean, intuitive, and highly functional dashboard that enables efficient monitoring of macroeconomic trends, relative country performance, and directional macro themes. The dashboard must include a dedicated trend analysis section where key macroeconomic indicators can be visualized as time-series line charts. This section should allow clear monitoring of directional trends over time. At minimum, this should include line charts for: • Inflation (CPI and/or Core CPI) • GDP (QoQ) • Central bank policy interest rates • 2-year government bond yields The trend section should: • Display historical time series with clean, professional formatting • Allow country selection (or comparison between multiple countries) • Clearly show directional trends, turning points, and relative positioning • Automatically update as new data becomes available The ideal candidate will have a strong background in macroeconomics, finance, and data visualization, with proven experience building professional dashboards using Excel, Power BI, Tableau, Python, or similar tools. Experience sourcing, structuring, and maintaining economic datasets from reliable providers such as FRED, central banks, statistical agencies, OECD, or commercial providers is highly desirable. The dashboard should be designed for reliability, clarity, and easy ongoing updates. Key responsibilities include designing the dashboard architecture, sourcing and integrating economic data, structuring the backend data framework, building clear and professional visualizations, and ensuring the system can be efficiently updated as new economic data is released. Screening Questions (required): 1. Briefly explain how you would approach designing and building this macroeconomic dashboard. What tools and structure would you use? 2. What data sources would you use to obtain reliable macroeconomic data for the listed countries? 3. How would you structure the data and dashboard to allow efficient ongoing updates? 4. Which platform would you recommend for this project (Excel, Power BI, Tableau, Python, or hybrid), and why? 5. Please provide examples of similar financial, economic, or macro dashboards you have built.
7 days ago19 proposalsRemoteBuilding a Strong Global AI Team!
We are looking to collaborate with highly skilled professionals in AI development and related fields. Our goal is to build a powerful, multilingual, and experienced international team. We are especially seeking experts in translation, voice projects, transcription, and data collection. Our Projects:
3 days ago18 proposalsRemoteMultilingual audio data collection project
We are conducting a multilingual audio data collection project and are seeking native speakers from specific regions to participate. The project involves recording natural, high-quality voice samples in your native language and regional accent. We are currently looking for native speakers of English (Ireland, New Zealand, Scotland, South Africa, Wales, Singapore), German (Switzerland), Chinese (Hong Kong), and Cantonese (China, Hong Kong). Participants must be born and raised in the respective region to ensure authentic pronunciation and accent accuracy. This is a remote, freelance opportunity suitable for individuals with clear speech and access to a quiet recording environment.
25 days ago4 proposalsRemoteSenior BI & Analytics Consultant
Our travel-focused platform runs on data, yet we are still hitting a few roadblocks: reports occasionally contradict each other, web experiments take too long to interpret, and we know there is hidden opportunity to move our conversion rate even higher. I want a senior-level mind to step in, find the story inside our numbers, and turn it into dashboards, recommendations, and winning test results. Scope and priorities • First, tighten the data foundation. Using Snowflake and Oracle you will write advanced SQL and, through dbt models scheduled in Airflow/Astronomer, ensure every metric is trustworthy. Eliminating inaccurate data reporting is the immediate win. • Second, elevate visualization. You will rebuild our Tableau estate so executives see traffic analysis, KPIs, and conversion funnels at a glance, with lightning-fast load times. • Third, unlock experimentation insights. Our Optimizely programme is active but under-analysed; you will pair the raw results with Adobe Analytics to resolve current difficulties in A/B testing analysis and surface clear next steps for site performance and personalization. Key tools in play include Tableau, Adobe Analytics, Optimizely, Snowflake, Oracle, dbt, Airflow, and Git for version control; deep knowledge of each is essential. A background in e-commerce, travel, or hospitality analytics will help you understand our audience and move quickly. Deliverables 1. Clean, documented SQL/dbt models feeding a single source of truth. 2. At least three executive-ready Tableau dashboards displaying website performance, traffic, and conversion rate optimization metrics. 3. A repeatable Optimizely analysis framework (queries, templates, slides) that shortens the time from test completion to decision. 4. A concise narrative deck highlighting insights, wins, and next steps—data storytelling that lands with non-technical stakeholders. Please attach or link to previous Tableau dashboards or BI work that demonstrate similar challenges solved. I will shortlist and invite a brief technical chat to confirm fit before kickoff.
5 hours ago6 proposalsRemoteLong term YouTube short editor
Hi, I'm Karo. I have a small YouTube channel about science-based nutrition and “what I eat” content (karo gsr on YT). I also work as a strategy consultant, so I’m mostly unavailable during weekdays. I’m looking for one reliable editor who can take ownership of the editing long-term. ✅ What I’ll Provide: - Raw footage - Structure for each video (usually highlighting meals like breakfast, lunch, snacks, etc.) ✂️ What I Need From You: - Clean, simple editing - no flashy effects or transitions - Consistent quality and quick turnaround times - Attention to detail - this is absolutely crucial (e.g. no spelling mistakes, balanced audio, no awkward cuts). I should be able to review once and upload. Pls include the word cherries at the start of your application. - Proactive - you want to proactively make the videos better every time and suggest improvements - Well organised - since I have no time during the week, you need to ensure you have everything you need upfront. I prefer that we brief once, instead of back-and-forth communication throughout the week. To apply, please share examples of similar previous work, rate per video and approximate turnaround time for a 1-minute short. As part of the application, I will ask you to edit one trial short before deciding on a long-term collaboration.
13 days ago39 proposalsRemoteopportunity
Cloudflare Worker Developer for Internal IP Tracking Tool
We’re looking for a developer to build a small internal tool to capture visitor IP data from our own website for testing and evaluation. The goal is to reliably capture and store raw visit data so we can assess usefulness over the next few months. This is not a full product build. Scope Build a Cloudflare Worker endpoint Capture and store: Visitor IP address Timestamp Page URL / path Referrer UTM parameters (if present) Store data in Supabase (PostgreSQL) Provide a simple JavaScript snippet to embed on the website Compatible with a Duda-hosted website Consent-safe (script should only run after consent – compatible with Usercentrics) Basic security (CORS / origin validation) Clean, readable, documented code Out of scope Dashboards or user interface CRM integrations IP-to-company enrichment This is a lean internal test build, not a product or platform. Deliverables Cloudflare Worker source code + deployment steps Supabase table schema Example embed script Short README explaining setup and data flow Process Before starting, we’ll have a short scoping discussion to confirm approach and assumptions so we’re aligned before work begins. Good fit if you Have hands-on experience with Cloudflare Workers Have worked with Supabase / PostgreSQL Understand how to correctly capture IPs behind proxies/CDNs Are comfortable building small, well-scoped internal tools
13 days ago27 proposalsRemoteAdministrative Assistant – Remote
We are seeking a reliable Administrative Assistant to provide remote support to our team. Key Responsibilities: Manage emails, calendar, and scheduling Prepare documents and reports Assist with data entry and research Support team communication and organization Skills & Requirements: Ability to work independently and meet deadlines Work Hours: Remote position Must be available during US business hours How to Apply: Submit your application directly through PeoplePerHour.
5 days ago28 proposalsRemoteDeveloper UK/US Financial News Screener (Python / APIs / Data)
PROJECT OVERVIEW I’m building a financial stock news screener focused initially on UK (AIM & Main Market) and then later US small-cap stocks. The system is designed to filter genuinely market-moving news from daily noise. This is an ongoing build, not a one-off task. I’m looking for one capable individual developer (not an agency) who can own the technical implementation and iterate with me as the logic evolves. Where necessary, I want the option to work side-by-side in person for short periods, so UK or Europe location matters. WHAT THE SYSTEM DOES (HIGH LEVEL) - Ingests real-time and scheduled stock news from multiple APIs (UK & US) - Parses and classifies news using firstly rules, secondly AI - Scores news for market impact (positive / negative / neutral) - Flags dilution, governance and risk signals - Outputs structured alerts and internal dashboards This is a logic-heavy, data-driven project, not a UI-first build. REQUIRED SKILLS (MUST-HAVE) Please do not apply unless you are comfortable with most of the following: - Python (primary language) - Working with REST APIs (news, market data, etc.) - Data parsing, filtering and scoring logic - SQL or NoSQL databases (Postgres, MongoDB or similar) - Clean, readable, maintainable code - Git / version control - Comfortable discussing system design and trade-offs Experience with financial data, trading tools, or market news is a major plus. NICE TO HAVE (NOT ESSENTIAL) - Experience with financial markets, trading, RNS or SEC filings - AI / LLM-based text classification - Elasticsearch or similar search tools - AWS or cloud deployment - Previous SaaS or data-platform builds LOCATION REQUIREMENT (IMPORTANT) You must be based in or near one of the following: - UK - Spain - Western or Southern Europe (easy travel to UK or Spain) Due to the likely evolving nature of the project I want the option to meet in person and potentially work together for short, focused periods if required. Please clearly state your location when applying. ENGAGEMENT TYPE - Individual freelancer only (no agencies) - Long-term potential if the fit is right - Paid hourly or milestone-based (open to discussion) HOW TO APPLY (VERY IMPORTANT) To avoid generic applications, please include: 1. A short description of a similar data-driven or API-heavy project you’ve worked on 2. Your primary tech stack 3. Your current location 4. Your hourly rate 5. Confirmation that you are open to occasional in-person collaboration Applications that ignore this will not be considered. WORKING STYLE I’m an equities/indices trader with 30 years experience, reasonably technical but not a developer. I value clear thinking, no nonsense honest communication, and someone who challenges bad ideas rather than blindly coding them.
20 days ago13 proposalsRemote