Filtre

Son aramalarım
Şuna göre filtrele:
Bütçe
ile
ile
ile
Tür
Beceri
Diller
    İş Durumu
    1,934 beautifulsoup iş bulundu

    ...tabanlı + gerekirse ML tabanlı pipeline oluşturmak. Kalite ölçütlerini (token sayısı, Türkçe oranı, örnekleme QC raporları) tanımlayıp düzenli raporlamak. NLP eğitim ekiplerine teslim edilecek final dataset (jsonl / parquet) ve dokümantasyon üretmek. Aradığımız kişi: Büyük ölçekli metin temizleme / data wrangling tecrübesi (en az 3 yıl). Python ekosistemine hâkim; pandas, multiprocessing, regex, BeautifulSoup, ftfy, spaCy/FastText tabanlı dil tespiti, vs. Tercihen Spark/Dask veya benzer dağıtık sistem deneyimi. Türkçe dil yapısına aşinalık ayrıca avantaj. Veriden rapor üretebilen (ör. Jupyter, Metabase) ve pipeline’ı otomasyona bağlayabilen (Airflow, Prefect) kişiler ter...

    €580 Average bid
    €580 Ortalama Teklif
    39 teklifler

    Freelance Yazılım Geliştirici Aranıyor Çalışma Türü: Uzaktan – Proje Bazlı Bütçe: Karşılıklı Belirlenecektir Aranan Nitelikler: Python veya JavaScript dillerinde ileri seviye bilgi sahibi Web tarama ve otomasyon araçlarında deneyimli (Selenium, Scrapy, BeautifulSoup vb.) Siber güvenlik araçlarına hâkim (OWASP ZAP, Burp Suite, Nikto vb.) Güvenlik açığı tespiti ve web uygulama güvenliği konusunda bilgi sahibi API geliştirme ve e-posta entegrasyonu yapabilen Temel seviyede yapay zekâ ile anomali tespiti bilgisi (tercih sebebi) İyi derecede yazılı iletişim becerisi Takım çalışmasına yatkın ve gizliliğe önem veren Tercih Sebepleri: Daha önce benzer projelerde görev almış ol...

    €25668 Average bid
    €25668 Ortalama Teklif
    14 teklifler

    ...Çekilen verilerin işlenerek sistemimize entegre edilmesi Web sitesinin 10 dilde hizmet verecek şekilde tasarlanması Otel fiyatlarının otomatik olarak %10 düşürülerek yayınlanması Kullanıcı dostu ve mobil uyumlu bir arayüz SEO uyumlu, hızlı ve güvenilir bir altyapı Otel bilgilerini düzenli olarak güncelleyen otomatik bir sistem Teknoloji Tercihleri: Backend: Python (Scrapy, Selenium, BeautifulSoup gibi web scraping kütüphaneleri), Node.js veya PHP Frontend: React.js, Vue.js veya Angular Veritabanı: PostgreSQL, MySQL veya MongoDB Dil Desteği: Çok dilli altyapı için gettext, i18n veya benzeri bir uluslararasılaştırma yöntemi Hosting & Deployment: AWS, Google Cloud veya DigitalOcean İş Teslim S&u...

    €422 Average bid
    €422 Ortalama Teklif
    21 teklifler
    Hava durumu
    Bitti left

    ... #### Gereksinimler: - Python - BeautifulSoup - Requests - Pandas - Matplotlib #### Adımlar: 1. **Veri Toplama:** - Web scraping ile bir hava durumu sitesinden veri çekmek. - Örneğin: `` 2. **Veri İşleme:** - Çekilen verileri anlamlandırmak ve Pandas DataFrame'ine dönüştürmek. - Gereksiz bilgileri temizlemek ve analiz için gerekli olanları seçmek. 3. **Veri Analizi:** - Günlük ortalama sıcaklık, nem oranı gibi bilgileri hesaplamak. - Zaman içinde hava durumu değişikliklerini gözlemlemek. 4. **Veri Görselleştirme:** - Matplotlib kullanarak verileri görselleştirmek. - Örneğin, belirli bir süre için sıcaklık değişim grafiği oluşturmak. #### Örnek K...

    €102 Average bid
    €102 Ortalama Teklif
    15 teklifler

    ...potential negotiation opportunities. DELIVERABLES 1. A working scraper (Python or similar tool). 2. A structured dataset stored in Excel, Google Sheets, or CSV format. 3. An Excel dashboard or structured sheet that allows quick filtering and analysis of the listings. 4. Clear instructions so the scraper can be run on my computer. TECHNICAL APPROACH Preferred tools include: Python, Selenium, BeautifulSoup, Scrapy, Apify, or similar scraping tools. The developer may propose the most stable solution. ACCEPTANCE CRITERIA • The scraper can run automatically once per day. • The dataset correctly tracks listings over time and avoids duplicates. • Price changes and days on market can be calculated. • The Excel file allows quick identification of potential o...

    €424 Average bid
    €424 Ortalama Teklif
    61 teklifler
    Scrape 7K Aliexpress Media
    6 gün left
    Onaylı

    I have around 7,000 Aliexpress products that I need fully harvested for content-creation purposes. From each listing I only require the official product photos and an...product URL or SKU will make downstream editing much easier for me. Deliverables: • Folder structure or archive segmented by product (one folder per listing). • Inside each folder: all JPEG images and any MP4 videos found. • A simple CSV mapping product URL → asset file names so I can trace anything quickly. I’m happy for you to choose the most efficient tooling—Python with Selenium, BeautifulSoup, or similar headless solutions are fine—as long as the final package is complete and safely transferable via cloud link. Let me know your estimated turnaround time and any clarifi...

    €28 Average bid
    €28 Ortalama Teklif
    22 teklifler

    I need a reliable, one-time scrape of roughly 300 public education websites that all share a similar page structure. Each site lists between 5 and 200 staff contacts (average ≈50). For every pers...URL where the data appears All pages are publicly accessible—no authentication hurdles—so the script can run headless without session handling. Please deliver: 1. A consolidated Excel file (.xlsx) containing every contact, with clear column headers and either a “Site” column or separate tabs—whichever keeps the data easiest to filter. 2. The scraper’s source code (Python with Scrapy, BeautifulSoup, or similar is fine) plus a brief README so I can rerun or tweak it later. Clean, deduplicated results and accurate mapping of each field to it...

    €129 Average bid
    €129 Ortalama Teklif
    101 teklifler

    ...professional-service businesses in major cities only. For every advertiser it detects, the CSV must list business name, Country, city, street address, email, phone, primary contact, and an estimated PPC budget. On top of that, the scrape should extract key campaign insights: • Budget • Keywords used • Ad placements A straightforward command-line script in Python is fine—Selenium, Scrapy, BeautifulSoup, SerpAPI, or the Google Ads API can all be leveraged so long as the solution stays within Google’s terms and reliably handles captchas, rotating proxies, and rate limits. Batch processing (multiple niches or cities fed from a text file) and clear logging are a plus. Deliverables • Fully commented source code • Setup/usage guide ...

    €175 Average bid
    €175 Ortalama Teklif
    22 teklifler

    ...property type, sqft, lot size, occupancy (if available), county/state, source URL, status (upcoming/active/sold/postponed/canceled), priority score (1–100), alert tier. Output Structured JSON via webhook to our Supabase REST API (we provide schema + credentials). Alert emails as formatted HTML via SendGrid/SES. No duplicates — dedup by address + parcel ID. Technical Requirements Python (Scrapy, BeautifulSoup, Selenium) and/or JavaScript (Puppeteer, Playwright) API integration (Supabase REST API, ATTOM Data API) Anti-bot handling: CAPTCHA solving, IP rotation, proxy management Government website scraping experience (fragile, inconsistent sites) PDF parsing and OCR Email automation (formatted HTML alerts) Scheduling/orchestration (cron, n8n, Airflow, or similar) Error ...

    €266 Average bid
    €266 Ortalama Teklif
    22 teklifler

    ...professional-service businesses in major cities only. For every advertiser it detects, the CSV must list business name, Country, city, street address, email, phone, primary contact, and an estimated PPC budget. On top of that, the scrape should extract key campaign insights: • Budget • Keywords used • Ad placements A straightforward command-line script in Python is fine—Selenium, Scrapy, BeautifulSoup, SerpAPI, or the Google Ads API can all be leveraged so long as the solution stays within Google’s terms and reliably handles captchas, rotating proxies, and rate limits. Batch processing (multiple niches or cities fed from a text file) and clear logging are a plus. Deliverables • Fully commented source code • Setup/usage guide ...

    €97 Average bid
    €97 Ortalama Teklif
    22 teklifler

    ...and logging — If a circuit scrape fails or returns no results, log the error with timestamp and circuit ID. Do not silently skip circuits. Optionally support email or webhook notification on failure. Scheduling-ready — The tool should run headlessly from the command line and be schedulable via cron or Windows Task Scheduler without manual intervention. Tech Stack Preferences: Python 3.x, BeautifulSoup or Playwright (for JavaScript-rendered pages), pandas for CSV output. Deliverable should include a and brief setup documentation. Deliverables: Working Python application with all 20 circuits implemented External config file for all circuit URLs and scraping strategies Sample output CSV matching the provided format Change-detection diff report on each run README w...

    €157 Average bid
    €157 Ortalama Teklif
    58 teklifler

    ...publicly visible customer review talking about Bayer from social-media sources—right now the focus is on Goole. The crawler should pull the full review text, star rating (or reaction score, if available), reviewer name or handle, date, and the direct URL to each post. Please build it so I can run it on demand, ideally from a simple command line or Jupyter notebook. Python with requests / BeautifulSoup, Selenium, or Scrapy is fine; if you prefer another stack, let me know why it would be a better fit. Deliverables • Clean, well-commented source code • One sample export in CSV or JSON showing at least 100 live reviews • A short README explaining environment setup, run instructions, and how to alter search terms or date ranges I’m only conce...

    €19 / hr Average bid
    €19 / hr Ortalama Teklif
    131 teklifler

    ...ideal workflow is simple: • I enter or update the list of product links I care about. • The script pings those pages at a reasonable interval, detects the switch in stock status, and immediately triggers a push notification to my smartphone. A lightweight web-scraper with a clear, maintainable rule for that button-text change should be enough. If you prefer to use Python with requests/BeautifulSoup —or Playwright, Puppeteer, or another headless approach—feel free; the key is fast, accurate detection and rock-solid alerts. Please include brief setup instructions so I can run it on my own machine or a small VPS and add or remove SKUs whenever I need. Deliverables 1. Source code with clear comments. 2. Configuration file or simple UI for adding prod...

    €134 Average bid
    €134 Ortalama Teklif
    85 teklifler

    ...stock options and awards, bonuses / incentive pay, plus any other compensation figures that appear in the summary or grants tables. Because the data are scattered in both narrative text blocks and embedded HTML tables, a purely scripted scrape misses too much, while a purely manual effort would be too slow. I’m therefore looking for a balanced workflow that blends solid Python-based parsing (BeautifulSoup, pandas, regex, maybe an LLM call for tricky passages) with targeted human review to catch formatting quirks and footnotes. Deliverables • A single CSV or Excel file where each row is a firm-year filing and each column holds one of the compensation items above, clearly labeled. • A short read-me describing the extraction logic, any LLM prompts used, and the...

    €1052 Average bid
    €1052 Ortalama Teklif
    146 teklifler

    ...delivered in a single, well-structured Excel file (.xlsx) with clear column headers and no duplicate rows. A short “read-me” tab or text file that explains any data cleaning or assumptions you had to make will also be appreciated. Because this is a one-off job, efficient turnaround is important to me. Please outline: 1. the approach and tools you will use (e.g., Python, Scrapy, Selenium, BeautifulSoup, Playwright, etc.) while respecting the site’s pagination and anti-bot measures; 2. the estimated time you need from award to delivery; 3. a realistic fixed price for the full scrape, including any post-processing needed to ensure clean, accurate data. If you can optionally supply the scraping script as part of the hand-off, note that in your proposal&mda...

    €144 Average bid
    €144 Ortalama Teklif
    68 teklifler

    I need a Python-based scraper that pulls complete c...useful is missed. Here’s what matters to me: • Reliability – the code must navigate pagination, work around basic anti-bot measures (rotating user-agents / respectful delays), and throw clear errors if the site layout changes. • Clean output – save to CSV or an SQLite database with consistent column names, ready for later analysis. You’re free to choose libraries you trust (requests, BeautifulSoup, Selenium, Scrapy, Playwright, etc.); just document any setup steps and keep third-party dependencies to a minimum. Deliverable: git-ready project folder containing the scraper, a brief README with run and schedule instructions, and a sample output file generated from at least a few live pages ...

    €31 Average bid
    €31 Ortalama Teklif
    37 teklifler

    I’m putting together a master spreadsheet of Australian businesses and need an organised, accurate data-gatherer to make it happen. ...spreadsheet (Excel or Google Sheets) with consistent formatting and no duplicates. Acceptance criteria • At least 95 % of golf clubs identified nationwide, verified by cross-checking multiple directories. • Phone numbers in Australian format, one row per business. • No duplicate entries and no blank mandatory fields. If you work comfortably with tools like Excel, Google Sheets, Python, BeautifulSoup or similar, you’ll fly through this. Data accuracy is more important than sheer speed, so double-checking entries is essential. Let me know how you plan to tackle the initial golf-club list and how quickly you can scale ...

    €262 Average bid
    €262 Ortalama Teklif
    104 teklifler

    ...property type, sqft, lot size, occupancy (if available), county/state, source URL, status (upcoming/active/sold/postponed/canceled), priority score (1–100), alert tier. Output Structured JSON via webhook to our Supabase REST API (we provide schema + credentials). Alert emails as formatted HTML via SendGrid/SES. No duplicates — dedup by address + parcel ID. Technical Requirements Python (Scrapy, BeautifulSoup, Selenium) and/or JavaScript (Puppeteer, Playwright) API integration (Supabase REST API, ATTOM Data API) Anti-bot handling: CAPTCHA solving, IP rotation, proxy management Government website scraping experience (fragile, inconsistent sites) PDF parsing and OCR Email automation (formatted HTML alerts) Scheduling/orchestration (cron, n8n, Airflow, or similar) Error ...

    €17 / hr Average bid
    €17 / hr Ortalama Teklif
    84 teklifler

    ...more APIs (for example OpenCorporates, Clearbit or any better suggestion you have). 3. Store results in a structured format (CSV plus an SQLite or Postgres option). 4. Offer a simple command-line trigger as well as a callable function so I can integrate it into larger workflows later. 5. Log activity and errors clearly. Tech stack Python 3.x with common libraries such as Requests, BeautifulSoup or Scrapy, Pandas and an ORM (SQLAlchemy is fine). If Selenium or Playwright is unavoidable for dynamic pages, please factor that in. Acceptance criteria • Full source code with virtual-env requirements file. • Sample run that fetches at least 30 real company records, shows enrichment working and saves the combined dataset. • README explaining setup, usage an...

    €192 Average bid
    €192 Ortalama Teklif
    16 teklifler

    ...“address eligibility” checker, and I need that information pulled automatically rather than re-typing each location by hand. Your task is to build and run a scraper that goes through the same steps the public tool requires, captures the eligibility result for every address I supply, and returns the full set in a clean Spreadsheet (Excel or CSV is fine). A repeatable script—Python with requests / BeautifulSoup or Selenium, or any language you are comfortable with—is preferred so I can rerun it later when the list of addresses grows. Handle captchas or session cookies if the site uses them, and respect polite crawling speeds to avoid blocking. Deliverables • The complete spreadsheet containing the eligibility outcome for every input address • ...

    €244 Average bid
    €244 Ortalama Teklif
    44 teklifler

    ...Companies) State Company Type (Private Limited / LLP / OPC / Public Limited) Authorized Capital (if available) Registered Office Address (if available) The system should: Run automatically daily (cron / scheduler) Avoid duplicate records Export data to Excel / CSV / Google Sheets / API endpoint Handle captcha or dynamic website structure if applicable Preferred technologies: Python (BeautifulSoup / Scrapy / Selenium) Node.js Puppeteer / Playwright Any robust scraping framework Deliverables: Fully working scraper or API system Source code Documentation for running the script Optional: Dashboard or automated email delivery of daily data Additional Preferred Features (Bonus): Historical data scraping Cloud deployment (AWS / DigitalOcean / VPS) API endpoint to ...

    €87 Average bid
    €87 Ortalama Teklif
    30 teklifler

    ...dynamic JavaScript-rendered pages. Experience with data parsing and structured data storage. Ability to implement error handling and logging. Deliverables: Fully functional scraping script or application. Clean, well-structured dataset. Documentation explaining how to run and maintain the scraper. Optional: automated scheduling or update mechanism. Preferred Skills: Python (Scrapy, Selenium, BeautifulSoup) or Node.js (Playwright, Puppeteer). Experience scraping large datasets. Experience with MongoDB or similar databases is a plus. Project Size: Medium to large. Please Include in Your Proposal: Your experience with similar scraping projects. Technologies you would use. Estimated timeline. Examples of previous work. We are looking for someone reliable who can build a scalab...

    €60 Average bid
    €60 Ortalama Teklif
    31 teklifler

    ...more APIs (for example OpenCorporates, Clearbit or any better suggestion you have). 3. Store results in a structured format (CSV plus an SQLite or Postgres option). 4. Offer a simple command-line trigger as well as a callable function so I can integrate it into larger workflows later. 5. Log activity and errors clearly. Tech stack Python 3.x with common libraries such as Requests, BeautifulSoup or Scrapy, Pandas and an ORM (SQLAlchemy is fine). If Selenium or Playwright is unavoidable for dynamic pages, please factor that in. Acceptance criteria • Full source code with virtual-env requirements file. • Sample run that fetches at least 30 real company records, shows enrichment working and saves the combined dataset. • README explaining setup, usage an...

    €214 Average bid
    €214 Ortalama Teklif
    23 teklifler

    ...substitutions, shots on goal, shots off target, etc. with times will help • normalise club names so they are consistent across all leagues, and • write everything into a single, tidy CSV file that can be appended with new match days at the push of a button or a scheduled task. I plan to run the scraper on my own machine, so please code in a language with minimal setup friction—Python with requests/BeautifulSoup or a headless browser such as Selenium/Playwright is ideal, but I am open to alternatives if they are well-documented. Acceptance is straightforward: I should be able to clone or unzip your project, install any listed dependencies, point to a configuration file (or run a default), and receive clean CSV output covering at least the current and previous s...

    €452 Average bid
    €452 Ortalama Teklif
    74 teklifler

    I have a curated list of URLs that I need parsed for their readable text only—no images, no embedded links, no HTML clutter. Once the crawl is finished, I want the raw text returned as straightforward .txt files, one file per source page. You are free to code in Python, Bash, Node, or any stack you prefer; common libraries such as BeautifulSoup, Scrapy, Selenium, or Playwright are fine so long as the final output meets the spec. Please respect , set reasonable delays between requests, and keep the natural order of the text exactly as it appears on the page. Deliverables • An executable script (with a brief README) so I can rerun the scrape in the future. • A zipped folder containing the plain-text results, clearly named after each URL. Let me know your estimat...

    €105 Average bid
    €105 Ortalama Teklif
    104 teklifler

    Regular comprehensive snapshot. There are 3,000 products. 20 columns for each product. Page by page. I’m looking for a repeatable, fully automated workflow. A Python-based stack (Scrapy, BeautifulSoup, Selenium, Playwright, or an equivalent you prefer). Robustness is key: the crawler should cope with pagination, JavaScript-rendered. Clear, well-commented code is part of the deliverable so my team can review and rerun it internally. Each quarterly hand-off must include: • Cleaned CSV or JSON containing the structured product records • The raw HTML or a compressed WARC snapshot for auditing • The executable script(s) plus a brief change log highlighting any site-structure updates you handled Please outline your proposed tool chain, an example of a large sc...

    €164 Average bid
    €164 Ortalama Teklif
    118 teklifler

    ...The scraper should capture: • Email addresses • Phone numbers • Mailing addresses • LinkedIn profile link • Location (city / state / country) • First and last name • Occupation / job title • Company name • Company website A well-structured CSV or Excel file is the preferred output, with each field in its own column. I am comfortable with your choice of tech—Python with BeautifulSoup, Scrapy, or Selenium are all fine—as long as the script runs reliably and respects and rate limits where required. Accuracy is more important than speed; every lead must be validated for correct formatting before delivery. Acceptance criteria 1. Script files and brief setup instructions are supplied. 2. Sample run on ...

    €206 Average bid
    €206 Ortalama Teklif
    31 teklifler

    ...Short description • Current stock option shown to shoppers (Delivery, Shipping, Pickup, etc.) The CSV has to be clean enough for immediate import into my market-analysis pipeline, so please normalize currency formats, trim whitespace and avoid duplicates. If either site relies on JavaScript for rendering, feel free to lean on Selenium or Playwright; otherwise a straightforward Python Scrapy/BeautifulSoup approach is fine. Please build the spider(s), run a full crawl, and hand over: 1. The finished CSV. 2. The runnable script or notebook with brief setup notes so I can refresh the data later. Accuracy and completeness are more important than speed. Let me know your preferred approach and an estimated turnaround once you’ve had a quick look at both URLs (I&r...

    €103 Average bid
    €103 Ortalama Teklif
    92 teklifler

    ...continuously without being blocked, parse every new listing, and filter it against a configurable set of electronics keywords. • Extract and store the Price and Condition fields so I can track changes and avoid duplicates. • Push an instant notification (email, SMS, or Slack—whichever you prefer to wire up) each time a fresh match is found. I’m comfortable with a Python 3 stack—think Requests/BeautifulSoup or Scrapy, perhaps headless Selenium for dynamic pages—so long as the code is clean, well-commented, and easy for me to extend. A simple settings file or ENV variables for keywords, polling interval, and notification credentials will be perfect. Deliverables will be the complete source, a quick README, and brief hand-off call showing ...

    €159 Average bid
    €159 Ortalama Teklif
    95 teklifler

    ...Key Responsibilities & Technical Stack You will be expected to automate the following "Trust Signal" engines using Python: • Geotagging Engine: Use piexif/Pillow to inject precise Lat/Long metadata into provided job photos. • GBP Activity Bot: Use google-api-python-client to schedule 2-3 weekly posts with AI-generated, location-specific captions. • Citation Gap Scraper: Build a scraper using BeautifulSoup/Selenium to identify and replicate top competitor directory listings (Yelp, Bing, etc.). • GHL Integration: Connect verified profiles to High-Level (GHL) and set up Webhook Alerts via Telegram/SMS for leads not replied to within 5 minutes. • Hyper-Local Injection: Scrape nearby landmarks (stadiums, parks, etc.) to inject into the "Serv...

    €414 Average bid
    €414 Ortalama Teklif
    140 teklifler

    ...be responsible for: - Analyzing government legal databases and regulatory websites - Building web scraping systems and crawlers - Automatically downloading legal documents and PDF files - Extracting metadata and source URLs - Organizing collected data into structured datasets - Creating an automated data collection pipeline TECHNICAL SKILLS (PREFERRED) - Python - Web Scraping - Selenium - BeautifulSoup - Scrapy - API integration - Data extraction automation - Data pipeline development DELIVERABLES The final deliverables should include: 1. Web scraping scripts or crawler system 2. Automated legal data collection pipeline 3. Downloaded legal document datasets (PDF files and documents) 4. Structured dataset including metadata and source URLs 5. Organized storage structure for ...

    €47 Average bid
    €47 Ortalama Teklif
    39 teklifler

    ...editing, and final posting with lead automation—must run without manual intervention. A multi-agent architecture is ideal, so feel free to leverage LangChain, CrewAI, AutoGPT, or any comparable framework that lets independent agents pass tasks between one another. Think of one agent focused on Google Keyword Planner or similar APIs, another handling SERP scraping with Python, Selenium, Scrapy or BeautifulSoup, a third using OpenAI or Stable Diffusion for media re-editing, and a final agent pushing the content live through my CMS and hooking into the CRM or email platform for lead collection. Deliverables • A reproducible, documented workflow (scripts, agent prompts, and environment files) • An installer or Docker setup that lets me spin the system up on my ow...

    €192 Average bid
    €192 Ortalama Teklif
    36 teklifler

    ...(e.g. “Plastic processing”) • Full address • Email • Phone number • Website • Number of employees • Year of foundation I will launch these requests on an ad-hoc basis, sometimes narrowing the crawl by geography, line of business or employee count, so the script or workflow should let me set those filters easily before each run. I’m happy with any stack you prefer—Python with BeautifulSoup/Scrapy, Playwright, Selenium, or a headless browser—so long as the final output lands in a well-structured .xlsx file with no missing fields, duplicate rows or formatting issues. Please let me know: • Your proposed approach to bypass pagination and collect the hidden email/phone fields shown on the profile pages....

    €134 Average bid
    €134 Ortalama Teklif
    141 teklifler

    ...Berkshire and London UK. We will pay per 2k list of the industries we will send upon acceptance. Email addresses must not be role based or trip any spam traps. I need a one-time extraction of verified email addresses from reputable online business directories. No other data fields are required—just the clean list of emails. Please choose whatever approach you prefer—Python with Scrapy/BeautifulSoup, browser automation with Selenium, or a similar tool chain—as long as the result is accurate and the scraping respects each site’s terms of service and rate limits. Deliverable • A CSV or XLSX file containing every unique email address you capture, de-duplicated and ready for import. Acceptance criteria • All emails must come exclusivel...

    €22 Average bid
    €22 Ortalama Teklif
    59 teklifler

    **Project Title** Build SaaS dashboard for legal-tech platform (Python FastAPI + React) **Project Overview** I am building a legal-tech platform called **Virtual Clerk** that automatically tracks court cases and alerts lawyers when hearing dates change. The backend scraping engine already exists. It uses: * Python * BeautifulSoup / Requests * Background workers * Modular court scrapers The system currently scrapes multiple courts and generates alerts and daily digests. What I need now is to turn this into a **usable web SaaS product for lawyers**. --- **Tech Stack** Backend: * Python * FastAPI Database: * PostgreSQL Workers / Queue: * Redis / Celery (or similar) Frontend: * React or Deployment: * Docker * VPS (DigitalOcean / AWS) --- **Required Features** 1. **U...

    €1134 Average bid
    €1134 Ortalama Teklif
    102 teklifler

    ...export results to CSV. My ideal flow looks like this: enter or schedule the URLs, run or auto-run the scraper, watch progress logs, and then immediately view fresh stats inside the dashboard—no command-line work once everything is deployed. If any source changes its HTML, the scraper should fail gracefully and flag the issue in the UI so I can react quickly. Tech stack is flexible; Python with BeautifulSoup / Scrapy on the back end and a simple Flask or Node/Express interface would be perfect, but I’m open to alternatives if they shorten development time or improve stability. Clean, well-commented code and clear setup docs are essential because I’ll be hosting the solution on my own VPS. Acceptance criteria • A working scraper module for each of the fo...

    €70 Average bid
    €70 Ortalama Teklif
    28 teklifler

    ...authenticated sessions without manual intervention. Core expectations • Extract and store: product titles/descriptions, all associated images, SKU codes, and current prices. • Export: structured CSV or JSON for data, separate folder (or S3 bucket) for images, with clear file naming that links each image back to its SKU. • Tech stack: Python with libraries such as Scrapy, Playwright/Selenium, BeautifulSoup, or a comparable approach—whatever you can prove is most efficient and resilient. Basic computer-vision or OCR hooks are welcome if they improve image handling. • Reliability: graceful error handling, automatic retries, and a simple log file so I can trace any failed requests. • Modularity: the list of target domains should live in a c...

    €422 Average bid
    €422 Ortalama Teklif
    224 teklifler

    ...but expect the usual high-traffic areas (Electronics, Real Estate, Automobiles) to be included. Data points • Vendor name • All available contact details (phone, email, WhatsApp, or any other channel the site exposes) Delivery You’ll give me: 1. The Excel file, one sheet per category, neatly labeled and deduplicated. 2. The script or notebook you used (Python with requests / BeautifulSoup, Scrapy, or a comparable tool is fine) so I can rerun the scrape later. Quality checklist (acceptance criteria) • Every row contains at least a vendor name plus one working contact detail. • No obvious duplicates. • File opens in current Excel without conversion warnings. If captcha handling, rate limiting, or hidden APIs are involved, f...

    €115 Average bid
    €115 Ortalama Teklif
    48 teklifler

    ...started throwing runtime errors before any data is returned. The pages themselves load fine in a browser, so the problem is clearly within my code or its dependencies. Here’s what you can expect from me: a zipped folder containing the current Python script, the list of target URLs, and a copy of the last error stack trace. I’ll also let you know which Python version and libraries (requests, BeautifulSoup, etc.) are in use so you can replicate the issue quickly. What I need from you: • Diagnose the exact cause of the errors • Deliver a clean, well-commented fix that reliably fetches all required text from each page • Briefly outline any library or environment changes I must make to keep the scraper stable in the future Once the script runs end-t...

    €26 - €218
    Mühürlü
    €26 - €218
    304 teklifler

    ...source is still open—fresh survey responses, publicly available web data, or even existing transactional records are all on the table. What matters is that the data be captured ethically, documented thoroughly, and delivered in a format ready for statistical analysis. Here’s what I need from you: • A brief proposal outlining the collection strategy, recommended tools (e.g., Qualtrics, Python/BeautifulSoup, Selenium, SQL, Excel), and estimated timeline • The final dataset in CSV or equivalent, fully de-duplicated and consistently formatted • A concise data dictionary describing each variable • A short methodology note (1-2 pages) explaining how the data was obtained, any limitations, and relevant citations or permissions Data integrity and ...

    €10 / hr Average bid
    €10 / hr Ortalama Teklif
    24 teklifler

    I need every piece of publicly available data pulled from three separate websites and packaged so I can work with it offline. The moment the contract is awarded I’ll hand you the URLs; from there I expect you to spin up an automated solution—Python with Scrapy, BeautifulSoup, Selenium, or a comparable stack—that captures all records, follows pagination, and handles any lazy-loaded content. Time is tight: the full job must be wrapped up within 3 calendar days. My total budget is fixed at US $300, so please be sure your approach fits comfortably inside that limit. Deliverables • A clean, well-structured file for each site (CSV or Excel, whichever you prefer) containing every scraped field. • The runnable scripts or notebooks you used, clearly commente...

    €348 Average bid
    €348 Ortalama Teklif
    176 teklifler

    I need a reliable way to collect product details from several e-commerce sites. I am open to either a fully automated scraper (Python, Selenium, BeautifulSoup, Playwright, etc.) or a well-structured manual process if that proves more stable for the target storefronts. Scope • Target pages: standard product listings and their individual detail pages on the chosen e-commerce sites. • Data fields: everything typically shown on a product page—title, SKU, description, images (URLs are fine), specifications, and the price shown at the moment of capture. • Output: please compile the final dataset in HTML or PDF, whichever reliably preserves the product information and any inline images. If your workflow first generates CSV/JSON/Excel and then converts, that is f...

    €42 Average bid
    €42 Ortalama Teklif
    52 teklifler

    I need the complete catalogue from another website duplicated in my Magento store. The task covers scraping every product name and full description, collecting all associated images, recording each SKU, and capturing the current stock level, then loading the ...correct path references. • A short hand-over note outlining the scraping steps so I can reproduce updates later. Acceptance criteria: at least 10 randomly chosen items will be spot-checked; all fields must match the source site, images must load, and no broken HTML may appear in the front end. Let me know your estimated timeline and the tools you intend to use—whether that’s Python with BeautifulSoup, Scrapy, Magento’s native import, or another reliable option—and share any similar catalogue m...

    €119 Average bid
    €119 Ortalama Teklif
    122 teklifler

    I need fresh, accurate leads pulled from business directories, regular websites, and the major social-media channels. For every record I expect four separate Excel columns—contact name, company name, email address, phone number, website/social media link, and physical address—so the sheet is ready for immediate upload into my CRM. You may use any stack you prefer (Python, Scrapy, Selenium, BeautifulSoup, Apify, browser automation, etc.); I just need the final hand-off in .xlsx. I do want the workflow documented briefly so future batches follow the same logic. Deliverables for each batch • Clean Excel file with no duplicates, one row per company • Columns: Contact Name I Company Name | Social Media link / website | Email Address | Phone Number | Address ...

    €106 Average bid
    €106 Ortalama Teklif
    113 teklifler

    ...once we agree). The job is straightforward: pull three data points for every product I specify—current price, stated delivery time, and offer type (Retail, Seller, or Seller but shipped by Retail). The site employs standard anti-bot measures, so rotating proxies, respectful delays, or headless browsing may be required to keep requests under the radar. Please build the script with Python—BeautifulSoup, Scrapy, Playwright, or any reliable stack is fine—as long as it can be rerun on my side without extra fees or licences. Deliverables • A single Excel workbook.  – One sheet per category or URL I provide  – Columns: Product identifier (ASIN/SKU/URL), Price, Delivery Time, Offer Type • The well-commented source code or note...

    €453 Average bid
    €453 Ortalama Teklif
    145 teklifler

    My small Python script scrapes a single website with BeautifulSoup and has reliably pulled the data I need for months. Yesterday the site pushed an update and the extraction part suddenly broke—no errors, just empty results. I have not changed anything on my end, so the issue is almost certainly new HTML structure, class names, or an extra layer of JavaScript that BeautifulSoup is now missing. Here is what I need: • Review the existing code, identify exactly why the selectors or parsing logic no longer match the updated markup. • Patch or refactor the script so it again captures every data field that used to be returned. • Keep the rest of the workflow (file writes, scheduling, etc.) untouched; only the scraping section needs attention. • Pro...

    €20 Average bid
    €20 Ortalama Teklif
    43 teklifler

    ...number • Business type • Full address • Star rating • Number of reviews Please return everything in a single, well-formatted Excel workbook so my sales team can filter and sort the data for outreach campaigns. The file should arrive free of duplicates and obvious errors, with each column clearly labelled and phone numbers in a consistent format. However you prefer to work—Python with BeautifulSoup or Selenium, Google Maps scraping libraries, or another reliable method—speed and accuracy are what matter most. If Google imposes any access limits along the way, you’ll need to handle proxies, delays, or captcha bypass so the sheet is complete. Deliverables are accepted once I can open the Excel file, scan a random sample of rows in G...

    €61 Average bid
    €61 Ortalama Teklif
    29 teklifler

    ...types and basic styling (date columns as dates, numbers as numbers, text wrapped where needed). Also send me the process and files so that I can test. Key points • 100 % coverage of all result pages on the site, not just a sample. • Output must be an Excel file with the cells pre-formatted for easy sorting and filtering. • Any scraping code or automation tool you build (Python + Selenium/BeautifulSoup, VBA, etc.) should be included so I can rerun the extraction when fresh results are posted. • If the site uses captchas, pagination or dynamic loading, handle those gracefully. • Keep the process compliant with the website’s terms of use. Deliverables 1. The completed, clean Excel workbook. 2. The runnable script or detailed steps you u...

    €66 Average bid
    €66 Ortalama Teklif
    22 teklifler

    I have a set of publicly available web pages and I need all visible, valid email addresses pulled from them quickly and accurately. The task is pure web-page scraping—no social media, no API work—focused solely on contact information, specifically emails. You are free to use Python with BeautifulSoup, Scrapy, Selenium, or any other language or framework you prefer, as long as the final result meets these simple acceptance criteria: • Every email captured comes from the target pages I supply. • Duplicates, malformed strings, and “noreply” style addresses are filtered out. • Results arrive in a clean CSV (column: Email) or Google Sheet, ready for me to import. Please include a brief note on your scraping approach and estimated turnaround t...

    €46 Average bid
    €46 Ortalama Teklif
    34 teklifler

    I need support working with the comments thread located at https://www.freelancer.com/groups/general-announcements/feed/3403...freelancer.com/groups/general-announcements/feed/340311/comments. Your task is to access that feed and deliver the full set of comments in a usable format so I can review them offline. Please ensure the output preserves each comment’s text, author name, and timestamp exactly as they appear on the page. Let me know: • the method or tool you would use to access the data (e.g., Python with BeautifulSoup, browser automation, API, etc.) • any limitations I should be aware of (rate-limits, login requirements, pagination) • how long you estimate the job will take A clean CSV or JSON file is perfect for delivery, but I’m open to an...

    €367 Average bid
    €367 Ortalama Teklif
    52 teklifler