Data scraping jobs
Please Sign Up or Login to see details.
...based on data scraped from a supplier website. The system should scrape products, stock availability, prices, images and product details from the supplier website, then publish them on my own website under my own branding, with my own higher prices. Main goal: scrape supplier products automatically sync stock and prices regularly apply my own markup rules display products on my own website accept customer orders create reorder workflow for supplier purchases Important: I do NOT need a copy of another company’s branding, logo, or copyrighted design. I need a custom-built website with similar functionality, powered by scraping-based product and stock synchronization. Required features: automatic scraping of products from supplier website scraping...
...and automation**. We are specifically looking for professionals who are experienced in **modern AI-driven marketing workflows** rather than traditional manual outreach methods. **Important Requirement – AI-Based Approach Only** This project must rely heavily on **AI tools, automation platforms, and scalable lead-generation systems**. Freelancers who rely primarily on **manual research, manual data entry, or manual outreach without automation** will not be considered. We expect the freelancer to propose and implement **automated, scalable marketing systems** using AI tools wherever possible. --- **Scope of Work** The freelancer will implement at least **three or four of the following marketing strategies using AI-powered tools**: **1. AI Directory & Startup Platf...
I’m putting together a master spreadsheet of Australian businesses and need an organised, accurate data-gatherer to make it happen. The first priority is every golf club in Australia; once those are complete, we’ll move straight on to other business categories so the file ultimately covers both golf clubs and a broad cross-section of Australian enterprises. What I need from you • Source publicly available information online—whether via targeted web scraping, reputable directories, or manual research—and capture each business’s name and primary phone number. • Include suburb, state and website where they’re easy to obtain so the list is genuinely useful, but the phone number is mandatory. • Deliver the results in a clea...
We would like a large number of pages extracted from a website. We have a detailed spec and exact deterministic plan on how you would achieve this. You would be running the initial bot to get previous pages and then running it ongoing and providing us with the code and data.
We are looking for a highly skilled AI and automation developer with experience in sports betting platforms and automated systems to help build and maintain an advanced betting automation solution. The ideal candidate should have experience working with sportsbook platforms, data scraping, AI prediction models, and automation tools to create a reliable system capable of processing large volumes of betting data and executing automated workflows. This is a freelance remote project with the potential for long-term collaboration.
No worries, that's way more workable. Here's the full post trimmed to fit under 10,000 characters: Nationwide Property Auction Web Scraping & Intelligent Alert System (Ongoing) About Us We're a commercial real estate investment firm that acquires distressed properties nationwide. We have the capital to close on any deal in the U.S. — our bottleneck is finding opportunities before competitors. We're building an automated system that monitors every property auction source in the country, filters against our criteria, and alerts us only on qualified deals. This is not a data dump project. We don't want spreadsheets with thousands of rows. We want a smart radar system that scans everything, filters ruthlessly, and only pings us when something m...
No scraping as they is not able to get the data. I need a current, easy-to-filter spreadsheet that captures every business operating in the four south-eastern Sydney suburbs of Malabar, Matraville, Little Bay, La Perouse and Maroubra (search in this order) For each entry, please research reliable public sources (official websites, Google Maps, directories, local chamber listings, etc.) and supply: • Business name • Full street address * Suburb name * Category * Email address * Link to contact form on website * Website • Facebook page link * Instagram page link • A short description of the main service or product offered * Source on how you find that info Make sure you don't miss any shops, cafe, restaurant, retail, etc. Accuracy matters...
...Monthly Bulk Member Adding Script (Telethon / Pyrogram) Description I am looking for an experienced Python developer who has previous experience working with Telegram automation using Telethon or Pyrogram. This is a recurring monthly task. I already have a list of Telegram usernames stored in a .txt file. The users are already filtered based on my requirements, so no additional filtering or scraping is needed. Your only responsibility will be adding these users to my Telegram group using a safe and reliable script. Task Details • I will send a .txt file containing around 5,000 Telegram usernames. • The usernames are already prepared and filtered. • Your task is only to add these users to my Telegram group using a script. • No group management, no moderat...
I need a robust, real-time scraping tool that monitors an accounting / legal / financial-services website and records every visitor’s key details the moment they arrive. The script must capture: • IP address • Name, phone and email • The exact Google or other search engine search keyword that led the user to the site • The specific page URL they land on and continue to browse • Any sign-up or form-fill actions performed • Geolocation (city/region/PIN Code) and device type Once each record is assembled it has to flow through my own verification endpoint first; if my custom API flags the data as unverified, the workflow should automatically fall back on a designated paid API to complete validation. All calls need to log their stat...
...spreadsheet (Excel or Google Sheets) containing at minimum: company name, decision-maker’s name and title, validated business email, website, city, and country. • Emails must be verified and ready for outreach; hard-bounce rate should stay under 5 %. • Priority is on owners, buyers, or purchasing managers—no catch-all or info@ addresses unless absolutely unavoidable. How you source the data is up to you—web scraping, LinkedIn Sales Navigator, trade-show directories, or specialized B2B databases are all fine—so long as the final list is accurate, up-to-date, and complies with CAN-SPAM/GDPR requirements. If you’ve built similar lead lists for jewelry, fashion, or luxury wholesalers before, that’s a plus. Please outline your app...
I need a Python-based solution that automatically gathers companies and shareholders data, pulls supplementary details via external APIs, and outputs a clean, unified dataset I can query at any time. Scope of the scrape • Sources: company websites, financial databases and relevant public records. • Website focus: company profiles, turnover figures and any available Demat / share-holding particulars. What the tool should do 1. Crawl or call the above sources, respecting and rate limits. 2. Parse the required fields, normalise names and IDs, then enrich each record through one or more APIs (for example OpenCorporates, Clearbit or any better suggestion you have). 3. Store results in a structured format (CSV plus an SQLite or Postgres option). 4. Offer a simpl...
...clean, ready-to-use list of prospects pulled directly from Google Maps. Please capture each company’s business name, physical address, phone number, and any email you can locate inside the listing so I can feed the data straight into my sales pipeline. Speed matters—ideally the first batch should land in my inbox as soon as possible, with the full file delivered shortly after. To keep the workflow smooth, deliver the results in a single Excel or CSV sheet, one row per record, free of duplicates and obvious errors. If you already have a proven method or custom tool for bulk scraping Google Maps without hitting quota limits, let me know. Accuracy and freshness of contact details will be the key success metric. Once you confirm you can hit these requirements, I&...
...through the website Or just scrape a given list of urls (from a txt-file) Key details • Sources: public-facing websites and shops (also with login using username:password) • Data type: text only—no images or binary files. • Output: one CSV per run, UTF-8 encoded, with a header row • should be able to read/exrtract data from !! various shops & websites !! -> generally i need a basic software + "plugins" for additional websites. This project should include ONE website (url upon request). Price for additional modules will be discussed before job is assigned • setting in application for multi-threading or scraping-delays to avoid IP-banning must be given I’m comfortable using a windows-application or a web-a...
I own a number of timeshare accommodation shares. Each year the timeshare management company releases a limited number of premium weeks to all timeshare owners. The release begins on a specific date at 8:00 am and typically all availab...button 5. on the book now screen, tick 2 boxes and click on submit 6. repeat the above process 0-7 times Essential Requirements: - No need for fancy user interface - Access a specified booking engine (e.g., , Airbnb, Expedia). - Complete accommodation bookings (type to be defined). - Programming language: Preferred JavaScript. Ideal Skills and Experience: - Experience with web scraping or API integration. - Knowledge of automation tools and scripting. - Familiarity with booking engines. - Good problem-solving skills and attention to detail.
...gateway integration (Stripe), OAuth flows, and working with third-party APIs. Ability to read and navigate an existing codebase - this is not a greenfield project. You'll be working within an established architecture and coding style. We value developers who can onboard themselves into existing code without needing extensive hand-holding. Web scraping/data collection - occasional tasks may involve building crawlers or scrapers (e.g., extracting structured data from external websites). Familiarity with HTTP clients, HTML parsing, and handling rate limits is a plus. Communication - clear, responsive communication in English is required. We use asynchronous communication and expect regular progress updates. Nice to Have Experience with Nuxt 3 / SSR Experie...
...high-end music distribution platform aimed at independent artists and record labels. I am looking for a talented web designer/developer to create a highly aesthetic, engaging, and 100% custom WordPress website. What this job is NOT: I am not looking for someone to install a pre-made theme (like Avada, Astra, etc.) or import a demo and change the text. Please do not apply if your workflow relies on scraping existing sites or using standard template blocks. What I am looking for: True Custom Design: You must be able to design a bespoke interface (preferably starting in Figma or Adobe XD) and translate that into a lightweight, high-performing WordPress site. Exceptional Aesthetics: The site needs to feel modern, sleek, and native to the music industry (think dark mode options, s...
We are looking for an experienced developer who can build an automated system to extract daily newly incorporated company data from the MCA (Ministry of Corporate Affairs) website – https://www.mca.gov.in. The system should automatically collect and deliver the list of companies incorporated each day in structured format (Excel / CSV / API / Database). Scope of Work: Develop a web scraping or API-based solution to extract daily incorporated company data from the MCA portal. The tool should automatically fetch newly incorporated companies every day. Data should include the following fields (minimum): CIN Company Name Date of Incorporation ROC (Registrar of Companies) State Company Type (Private Limited / LLP / OPC / Public Limited) Authorized Ca...
...build a robust web scraping solution capable of extracting structured data from a login-protected medical/drug repository website. The platform contains a large database of drug information (potentially hundreds of thousands to over a million pages). The scraper should be able to navigate through the website after login, systematically extract relevant drug data, and store it in a structured format. Scope of Work: Develop a scraper that can log into a protected website. Navigate through the drug repository pages. Extract structured information from each drug page. Handle pagination and large-scale crawling. Implement mechanisms to prevent crashes or interruptions during long scraping runs. Store extracted data in a structured format such as JSON, CSV...
...sellers. Ideal targets are doing roughly $50k–$500k per month in Beauty & Personal Care, Health Supplements, Pet Supplies, Home & Kitchen, or any other category prone to counterfeiting. Sellers may be based anywhere—US, India, Europe, Africa and beyond—so long as the contact data is accurate and legal to use. My outreach platform flags hard bounces, so please lean on reputable public data directories, Hunter-style email verification tools, and—when useful—social media profiles to keep the list clean. No gray-area scraping. Deliver the finished file in Google Sheets or CSV with these columns: • Seller / Store name • Main ASIN or clear category indicator • Verified email address (low bounce rate) • Pho...
...non-commercial accounts—and capture verified contact details rather than generic “info@” addresses whenever possible. To make the data truly useful, each row should include: • Influencer’s full name and Instagram handle • Public email address (primary) • Follower count and engagement rate (likes/comments % is fine) • Primary music genre or sub-niche they identify with • Country or city, if available • Source link or screenshot that confirms the email Accuracy matters more than sheer volume; I would rather have 500 solid entries than a bloated list that needs extensive cleaning. If you like using tools such as Phantombuster, , or manual scraping techniques, feel free—as long as the final file is relia...
I have a single website that lists venues and I need a clean spreadsheet pulled from it. Once we start, I will shar...spreadsheet pulled from it. Once we start, I will share the exact URL so you can inspect the structure before you begin. For every venue that appears on the site, I want these fields captured: • Venue name • Email address • Phone number • Full physical address Please scrape the entire catalogue—restaurants, event spaces, hotels or any other venue type the site includes—then deliver the data in CSV or Excel format with one row per venue and clearly labeled columns. I’m happy to answer any structural questions about the site up-front and will consider the job complete when the file imports without errors and sample check...
I need a reliable script that automatically pulls live product-inventory information from and returns the results in clean, well-structured JSON. Order details or shipment tracking are not in scope for now; the only data I care about is the current stock level and any related inventory fields that appear on each product listing. You’re free to choose the most stable approach—requests-based scraping, headless Selenium, or an API workaround—so long as it copes gracefully with dynamic content, anti-bot measures, and Mandarin text. The solution should run on a Linux VPS, include clear setup instructions, and be easy for me to schedule through cron once we decide on an optimal frequency. Deliverables • Python (or equivalent) source code with comments ...
...and video editing give you bonus points. Additionally, the role includes web scraping, email marketing, and tasks requiring excellent written English (beyond just AI). Key Requirements: Primary: - Proficiency in SEO, keyword research (SEMrush etc), link building and traffic generation - Social media management, design and content creation (blog and email) (primary focus) - Experience Figma and Adobe illustrator is a bonus (but not essential, just a bonus) - Strong proficiency in using generative AI for content, video, graphics and vibe coding - Strong written English for various tasks Secondary: - Experience in Claude code, n8n, sumopod, APIs etc (also not essential but a bonus) - Ability to perform web scraping tasks efficiently Ideal Skills and Experience: - Previou...
I need a Python-based solution that automatically gathers companies and shareholders data, pulls supplementary details via external APIs, and outputs a clean, unified dataset I can query at any time. Scope of the scrape • Sources: company websites, financial databases and relevant public records. • Website focus: company profiles, turnover figures and any available Demat / share-holding particulars. What the tool should do 1. Crawl or call the above sources, respecting and rate limits. 2. Parse the required fields, normalise names and IDs, then enrich each record through one or more APIs (for example OpenCorporates, Clearbit or any better suggestion you have). 3. Store results in a structured format (CSV plus an SQLite or Postgres option). 4. Offer a simpl...
I need a steady stream of qualified multifamily-building owners who are actively looking to cut operating expenses. My focus is very specific: owners motivated by energy-efficiency improvements to s...delivered in a clean spreadsheet ready for outreach. Acceptance criteria • Minimum 90 % contact accuracy (emails must pass validation). • Owners must control at least one building with 50 or more residential units. • Each lead’s interest in lowering operating costs through efficiency measures must be documented. If you already work with tools like ZoomInfo, Apollo, , or your own scraping/verification stack, tell me—speed and accuracy matter more than method. I’ll start with a small batch to confirm quality, then scale up quickly once we’re c...
I'm seeking a Python RPA expert specializing in computer vision-based (Offline) web scraping for a web search and document download project. You'll scrape sites, download/classify documents (e.g., public records via CV/NLP), design neural networks for extraction, and build scalable workflows. Patience for alpha testing, finetuning iterations is key. Key Requirements: • Pure Python RPA, with the core orchestrator, no third part tool. • Web navigation/ scraping with Selenium/Playwright: document download, classification, OCR/text extraction. • Build/train neural networks (e.g., CNNs for image doc classification). • NLP expertise with spaCy for entity extraction. • Computer vision using TensorFlow/OpenCV (offline Vision Libr...
I have a set of online databases that contain numerical figures I need captured with care and entered by hand into a clean, well-structured spreadsheet. No scraping tools or API hooks this time—I specifically want a diligent manual process to ensure every value is double-checked for accuracy and any anomalies are flagged before delivery. You’ll receive the site links, the exact fields to copy, and the target spreadsheet template (Excel or Google Sheets—whichever you prefer). Once the work is complete, I’ll review a sample to confirm format and precision before you proceed with the remainder of the entries. Final deliverable: the fully populated spreadsheet, clearly organized and free of transcription errors, ready for immediate analysis on my side.
I need a rock-solid n8n workflow that, whenever I trigger it, navigates through selected e-commerce sites and public business directories, captures every piece of business information that is publicly available, and stores it in a clean, query-ready format. The data I care about includes the business name, category or type, “about” text, founders’ names, any additional corporate details the site reveals, plus all images properly downloaded and tagged. I will be running various data-analysis models on the output, so accuracy, consistency, and tidy structuring are non-negotiable. The flow must: • Accept a list of target URLs and run on demand (no fixed schedule). • Respect and site rate limits while still remaining efficient. • Handle pag...
I need a reliable researcher who can compile a database of 1,000 Indian-based YouTube channels, centred on the “Data entry” niche, and capture a full-page screenshot of each channel’s “About” tab that clearly shows the Promotion/Business email address. Scope of work • Data collection and research: locate active Indian YouTubers whose content relates to data entry (tutorials, tips, freelancing, software demos, etc.). • Contact details: extract the publicly listed Promotion/Business email visible in the About section. • Screenshot capture: take a clear PNG or JPG screenshot of the About page for each channel so the email is readable. • Organise results: create a spreadsheet (Google Sheets or Excel) with Channel N...
...and video editing gives you bonus points. Additionally, the role includes web scraping, email marketing, and tasks requiring excellent written English (beyond just AI). Key Requirements: Primary: - Proficiency in SEO, keyword research (SEMrush etc), link building and traffic generation - Social media management, design and content creation (blog and email) (primary focus) - Experience Figma and Adobe illustrator is a bonus (but not essential, just a bonus) - Strong proficiency in using generative AI for content, video, graphics and vibe coding - Strong written English for various tasks Secondary: - Automation experience in n8n, sumopod, APIs etc (also not essential but a bonus) - Ability to perform web scraping tasks efficiently Ideal Skills and Experience: - Previous...
...sports-focused website. The job is purely about gathering the addresses—no additional analytics or marketing content required—so I can work with the raw data afterward. Please crawl every publicly accessible contact page on the site, capture each unique email, and return the final set in a simple spreadsheet (CSV or XLSX). Duplicates, role-based addresses that clearly repeat, and any non-functional strings should be removed before delivery. I am looking for: • Complete coverage of all contact pages on the site • A final file that lists each email once, with no formatting errors If anything on the site blocks automated scraping, let me know so we can decide whether to switch to a manual approach or adjust pacing. Once you send the spreadsh...
...that excite both of us. Whether it’s a handy automation script, a data-driven dashboard, a lightweight Flask or Django web app, a web-scraping utility, or even a small game, I’m happy to explore any direction—as long as it showcases clean, well-documented Python code. Because I do not have a strict deadline (No time limit), I prefer quality over speed. Take the time to think through the concept, architecture, and tech stack; then send me a Detailed project proposal that explains: • The core idea and its personal value • Key Python libraries or frameworks you plan to use (e.g., Pandas, Selenium, FastAPI, PyGame—whatever fits) • A rough work plan with milestones and an estimated timeframe • What you need from me (data...
...how you gather the data: scraping, APIs, or well-documented manual checks are all fine as long as the results are accurate. Deliverables • Spreadsheet (Excel or Google Sheets) listing: – Store URL and brand name – Estimated monthly traffic figure and data source (Similarweb, BuiltWith, etc.) – Primary product category – Proof that the Shopify Fraud Protection script is present (script URL, screenshot, or HTML snippet link) Acceptance criteria I’ll randomly sample about 10 % of the entries; any that fail the stated rules must be replaced. Clear, well-sourced data will secure fast approval and the possibility of follow-up work. Include a short note on your approach and, if possible, a mini sample of ...
I need a reliable scraping workflow that gathers both text and images from a set of public-facing websites and a collection of PDF files, then prepares that material into an excel file and stores images in a file where the image name is referenced on the excel document provided. This detail will be fed into my CMS and published onto our own site. For the web sources, the scraper should navigate through all relevant pages, capture the product details, text along with associated image(s), and return clean, structured output into the excel provided ready for ingestion into my CMS. The PDF portion is similar: extract full text and each embedded image from every document in the batch, preserving page order and basic layout indicators so I can re-render the content online. Accuracy in ...
hi, need someone to scrape some data. I need a google sheet with the name, email, phone and website of every company listed here https://www.freizeitmesse.de/ausstellerverzeichnis/#/suche/f=h-entity_orga;v_sg=0;v_fg=0;v_fpa=FUTURE Should be around 500 entries. Please share your rate and timeframe.
...and video editing gives you bonus points. Additionally, the role includes web scraping, email marketing, and tasks requiring excellent written English (beyond just AI). Key Requirements: Primary: - Proficiency in SEO, keyword research (SEMrush etc), link building and traffic generation - Social media management, design and content creation (blog and email) (primary focus) - Experience Figma and Adobe illustrator is a bonus (but not essential, just a bonus) - Strong proficiency in using generative AI for content, video, graphics and vibe coding - Strong written English for various tasks Secondary: - Automation experience in n8n, sumopod, APIs etc (also not essential but a bonus) - Ability to perform web scraping tasks efficiently Ideal Skills and Experience: - Previous...
...how you gather the data: scraping, APIs, or well-documented manual checks are all fine as long as the results are accurate. Deliverables • Spreadsheet (Excel or Google Sheets) listing: – Store URL and brand name – Estimated monthly traffic figure and data source (Similarweb, BuiltWith, etc.) – Primary product category – Proof that the Shopify Fraud Protection script is present (script URL, screenshot, or HTML snippet link) Acceptance criteria I’ll randomly sample about 10 % of the entries; any that fail the stated rules must be replaced. Clear, well-sourced data will secure fast approval and the possibility of follow-up work. Include a short note on your approach and, if possible, a mini sample of ...
...gather the data: scraping, APIs, or well-documented manual checks are all fine as long as the results are accurate. Deliverables • Spreadsheet (Excel or Google Sheets) listing: – Store URL and brand name – Estimated monthly traffic figure and data source (Similarweb, BuiltWith, etc.) – Primary product category – Proof that the Shopify Fraud Protection script is present (script URL, screenshot, or HTML snippet link) Acceptance criteria I’ll randomly sample about 10 % of the entries; any that fail the stated rules must be replaced. Clear, well-sourced data will secure fast approval and the possibility of follow-up work. Include a short note on your approach and, if possible, a mini samp...
...script) and handle the back-and-forth until the prospect agrees to a call. • Quotas & tracking: 60–80 quality DMs sent per day and a clear log of replies, follow-ups, and meetings secured. At week’s end I expect a short report summarizing outreach volume, positive responses, and meetings scheduled. • Tools: You may use IG Creator Studio, chat templates, basic CRM/Google Sheets, and any safe scraping or automation aids you already trust—as long as everything stays compliant with Instagram’s limits. • Voice & tone: Conversational, helpful, never spammy. Personalize each opener enough that it feels human. Success for me is a steady stream of calendared calls with qualified, medium-sized prospects. If you have ideas for refining the ...
...LangChain/LangGraph. Frontend/Dashboard: React.js or (Client & Admin Dashboards). Backend: Node.js or Python (FastAPI/Flask). Database: PostgreSQL or MongoDB with Vector Storage (Pinecone/Weaviate) for RAG. Lead Gen/Scraping: Apollo API, LinkedIn Automation, or custom headless scrapers. Communication: WhatsApp Business API (via Twilio/Cloud API), Botpress/ManyChat, and SendGrid. Document AI: AWS Textract, Google Document AI, or OCR.space. Voice AI: Vapi, ElevenLabs, or Bland AI. Scope of Work (Key Deliverables) 1. AI Lead Generation & Scoring Engine Automated scraping and lead capture from LinkedIn, business directories, and NRI forums. Automated outreach via Email, LinkedIn, and WhatsApp. AI Lead Scoring: Categorizing leads into Hot/Warm/Cold based on ...
I need a concise, well-researched article on a current technology topic delivered fast and error-free. The piece should read smoothly for a general audience yet include enough depth to show authority. Originality is critical—no AI scraping or recycled content—and the final copy must pass standard plagiarism checks. A clean structure with an engaging introduction, clear sub-headings, and a logical flow is expected. Please weave in relevant keywords naturally so the text is search-friendly without feeling forced. The tone should stay informative and neutral, avoiding jargon wherever possible. Deliverables • One polished technology article (around 800–1,000 words) in .docx or Google Doc format • A short meta title (max 60 characters) and meta descr...
...Likes and comments - Followers system - Groups (created by admin only) - Popular content sections MEDIA UPLOAD Users can upload: Photos: JPG, PNG, WEBP Videos: MP4, MOV, WEBM Admin configuration: - Maximum photo size - Maximum video size - Daily upload limits Video processing: - compression - conversion - automatic thumbnail generation Media protection: - watermark - temporary URLs - anti-scraping protection SUBSCRIPTION SYSTEM Subscriptions only (no individual paid content). Plans: FREE – Limited access BASIC – 4.99 €/month PREMIUM – 7.99 €/month VIP – 9.99 €/month PAYMENT METHODS - CCBill - PayPal - Crypto via NOWPayments - SMS payment (optional) PWA MOBILE APP The platform must work as a Progressive Web App. Compatible with: - An...
...script) and handle the back-and-forth until the prospect agrees to a call. • Quotas & tracking: 60–80 quality DMs sent per day and a clear log of replies, follow-ups, and meetings secured. At week’s end I expect a short report summarizing outreach volume, positive responses, and meetings scheduled. • Tools: You may use IG Creator Studio, chat templates, basic CRM/Google Sheets, and any safe scraping or automation aids you already trust—as long as everything stays compliant with Instagram’s limits. • Voice & tone: Conversational, helpful, never spammy. Personalize each opener enough that it feels human. Success for me is a steady stream of calendared calls with qualified, medium-sized prospects. If you have ideas for refining the ...
PROJECT TITLE Web Scraping Developer for Global Legal & Regulatory Data Collection PROJECT OVERVIEW We are looking for a developer who can build an automated system to collect legal and regulatory documents from multiple global sources. The goal is to create a scalable automated pipeline that can gather legal data across multiple jurisdictions and regulatory domains. DATA COLLECTION SCOPE The system will collect information related to: - Medical law and healthcare regulation - Medical advertising regulation - Corporate formation and company governance laws - Investment regulation (stocks, cryptocurrency, real estate) - Tax law and administrative tax rulings - Beauty and cosmetic regulation - Medical and cosmetic manufacturing compliance - Import and exp...
...* Complete **automated marketing system** * **Auto content generation with AI** * **Auto posting across multiple platforms** * **Lead collection system** * **Central dashboard or reporting** * Documentation explaining the system --- **Preferred Skills** We are looking for experts with experience in: * Growth hacking * AI marketing automation * Zapier / Make / n8n * Python automation * Web scraping * Social media automation * Lead generation systems * CRM integrations * API integrations --- **Important** Please include the following in your proposal: 1. Examples of **automation systems you have built** 2. Tools you plan to use 3. How you will automate **posting + lead capture** 4. Estimated time to build the system Only apply if you have experience building **AI-powered a...
...re-edits the visuals to match my branding guidelines, then publishes the finished piece to my website while triggering the appropriate lead-capture sequence. All three stages—keyword research and analysis, scraping plus in-house editing, and final posting with lead automation—must run without manual intervention. A multi-agent architecture is ideal, so feel free to leverage LangChain, CrewAI, AutoGPT, or any comparable framework that lets independent agents pass tasks between one another. Think of one agent focused on Google Keyword Planner or similar APIs, another handling SERP scraping with Python, Selenium, Scrapy or BeautifulSoup, a third using OpenAI or Stable Diffusion for media re-editing, and a final agent pushing the content live through my CMS and ho...
Want to automate how I track team moves and new/emerging PE firms. Web scraping PE firm team pages • Scrape c. 50–100 private equity firm “Team/People” pages. • Extract: firm name, person name, title/role, location (if available). • Export to Excel/CSV and make it easy to re-run weekly (script or small tool). • Ideally include a simple comparison versus previous runs to flag: new people, departures, and title changes. • Handle different page structures and basic anti-bot measures while respecting legal/ethical constraints.
...validation layer LLM Layer Invoked only after successful retrieval Rules must be enforced outside the model, in the orchestration layer. --- Infrastructure Possible stack: Python backend Vector database (Pinecone, Weaviate, Qdrant, etc.) Hetzner infrastructure API layer for query handling Content ingestion may come from: controlled exports PDFs structured feeds limited and controlled scraping if necessary --- Deliverables RAG backend architecture ingestion pipeline vector database configuration retrieval and validation logic API layer deployment instructions documentation --- Screening Question Please answer this question in your proposal: How would you implement a strict “no-retrieval = no-answer” rule in a production RAG architectur...