#python #ai #artificial_intelligence #automation #crawler #scrape #scraper #scraping #web_scraping #webautomation #webscraping
https://github.com/alirezamika/autoscraper
https://github.com/alirezamika/autoscraper
GitHub
GitHub - alirezamika/autoscraper: A Smart, Automatic, Fast and Lightweight Web Scraper for Python
A Smart, Automatic, Fast and Lightweight Web Scraper for Python - alirezamika/autoscraper
#html #node #scraper #video_downloader #youtube #youtube_downloader
https://github.com/fent/node-ytdl-core
https://github.com/fent/node-ytdl-core
GitHub
GitHub - fent/node-ytdl-core: YouTube video downloader in javascript.
YouTube video downloader in javascript. Contribute to fent/node-ytdl-core development by creating an account on GitHub.
#python #archive #downloader #gfycat #imgur #reddit #scraper
https://github.com/aliparlakci/bulk-downloader-for-reddit
https://github.com/aliparlakci/bulk-downloader-for-reddit
GitHub
GitHub - Serene-Arc/bulk-downloader-for-reddit: Downloads and archives content from reddit
Downloads and archives content from reddit. Contribute to Serene-Arc/bulk-downloader-for-reddit development by creating an account on GitHub.
#python #api #async #crawler #douyin #douyin_api #douyin_scraper #douyin_tiktok_api #douyin_tiktok_download #fastapi #no_watermark #online_parsing #python #pywebio #scraper #spider #tiktok #tiktok_api #tiktok_scraper #tiktok_signature #web_scraping
This tool, called "Douyin_TikTok_Download_API," allows you to easily download videos and data from Douyin (Chinese TikTok), TikTok, and Bilibili without watermarks. Here’s how it benefits you The tool is ready to use out of the box and supports API calls, online batch parsing, and downloading.
- **Versatile** Built with PyWebIO, FastAPI, and HTTPX for fast asynchronous data scraping.
- **Customizable** Provides comprehensive API documentation for easy integration into your projects.
- **iOS Integration**: Supports iOS shortcuts for quick video downloads without watermarks.
Overall, this tool simplifies the process of downloading content from these platforms while offering flexibility and high performance.
https://github.com/Evil0ctal/Douyin_TikTok_Download_API
This tool, called "Douyin_TikTok_Download_API," allows you to easily download videos and data from Douyin (Chinese TikTok), TikTok, and Bilibili without watermarks. Here’s how it benefits you The tool is ready to use out of the box and supports API calls, online batch parsing, and downloading.
- **Versatile** Built with PyWebIO, FastAPI, and HTTPX for fast asynchronous data scraping.
- **Customizable** Provides comprehensive API documentation for easy integration into your projects.
- **iOS Integration**: Supports iOS shortcuts for quick video downloads without watermarks.
Overall, this tool simplifies the process of downloading content from these platforms while offering flexibility and high performance.
https://github.com/Evil0ctal/Douyin_TikTok_Download_API
GitHub
GitHub - Evil0ctal/Douyin_TikTok_Download_API: 🚀「Douyin_TikTok_Download_API」是一个开箱即用的高性能异步抖音、快手、TikTok、Bilibili数据爬取工具,支持API调用,在线批量解析及下载。
🚀「Douyin_TikTok_Download_API」是一个开箱即用的高性能异步抖音、快手、TikTok、Bilibili数据爬取工具,支持API调用,在线批量解析及下载。 - Evil0ctal/Douyin_TikTok_Download_API
❤1
#javascript #batch_processing #batch_script #code_free #crawler #data_collection #frontend #gui #html #input_parameters #layman #parameters #robotics #rpa #scraper #spider #visual #visualization #visualprogramming #web #www
EasySpider is a free, code-free web crawler software that helps you collect data from websites easily. You can use it without writing any code, just by selecting the content you want to operate on a web page and following the prompts. It can also be run from the command line, making it easy to integrate into other systems. This software is very user-friendly and allows you to collect data from websites quickly, even if you don't know how to code. It also supports various proxy services and captcha solutions, making data collection more efficient. Using EasySpider can save you a lot of time and effort in collecting web data.
https://github.com/NaiboWang/EasySpider
EasySpider is a free, code-free web crawler software that helps you collect data from websites easily. You can use it without writing any code, just by selecting the content you want to operate on a web page and following the prompts. It can also be run from the command line, making it easy to integrate into other systems. This software is very user-friendly and allows you to collect data from websites quickly, even if you don't know how to code. It also supports various proxy services and captcha solutions, making data collection more efficient. Using EasySpider can save you a lot of time and effort in collecting web data.
https://github.com/NaiboWang/EasySpider
GitHub
GitHub - NaiboWang/EasySpider: A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:…
A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。 - NaiboWang/EasySpider
#typescript #api #automation #browser #browser_automation #low_code #maxun #no_code #playwright #robotic_process_automation #rpa #scraper #self_hosted #spreadsheet #web_automation #web_scraper #web_scraping #website_to_api
Maxun is a simple way to extract data from websites without coding. You can train a robot in just 2 minutes to automatically scrape web data for you. Here’s how it helps: You can create robots that capture lists, text, or screenshots from websites. These robots can run on their own, even handling pagination and scrolling, and can be scheduled to run at specific times. Maxun also integrates with tools like Google Sheets and offers a cloud version that manages anti-bot detection and proxy rotation, making it easy to extract data at scale. This saves you time and effort, allowing you to focus on other tasks while the robots do the work for you.
https://github.com/getmaxun/maxun
Maxun is a simple way to extract data from websites without coding. You can train a robot in just 2 minutes to automatically scrape web data for you. Here’s how it helps: You can create robots that capture lists, text, or screenshots from websites. These robots can run on their own, even handling pagination and scrolling, and can be scheduled to run at specific times. Maxun also integrates with tools like Google Sheets and offers a cloud version that manages anti-bot detection and proxy rotation, making it easy to extract data at scale. This saves you time and effort, allowing you to focus on other tasks while the robots do the work for you.
https://github.com/getmaxun/maxun
GitHub
GitHub - getmaxun/maxun: Turn any website into clean data pipelines & structured APIs in minutes!
Turn any website into clean data pipelines & structured APIs in minutes! - getmaxun/maxun
#python #agent #application_resume #artificial_intelligence #automate #automation #bot #chatgpt #chrome #gpt #human_resources #job #jobs #jobsearch #jobseeker #opeai #python #resume #scraper #scraping #selenium
AIHawk is an AI-powered job search assistant that helps you automate job applications, get personalized recommendations, and land your dream job faster. Here’s how it benefits you AIHawk automates the tedious process of filling out job application forms, saving you time and effort.
- **Personalized Recommendations** It generates customized resumes and cover letters to match the requirements of each job you apply for.
- **Community Support** The project is open-source and actively maintained by a team of contributors, ensuring it keeps getting better with user input.
Overall, AIHawk streamlines the job search process, making it more efficient and increasing your chances of finding the right job quickly.
https://github.com/feder-cr/Jobs_Applier_AI_Agent
AIHawk is an AI-powered job search assistant that helps you automate job applications, get personalized recommendations, and land your dream job faster. Here’s how it benefits you AIHawk automates the tedious process of filling out job application forms, saving you time and effort.
- **Personalized Recommendations** It generates customized resumes and cover letters to match the requirements of each job you apply for.
- **Community Support** The project is open-source and actively maintained by a team of contributors, ensuring it keeps getting better with user input.
Overall, AIHawk streamlines the job search process, making it more efficient and increasing your chances of finding the right job quickly.
https://github.com/feder-cr/Jobs_Applier_AI_Agent
GitHub
GitHub - feder-cr/Jobs_Applier_AI_Agent_AIHawk: AIHawk aims to easy job hunt process by automating the job application process.…
AIHawk aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in a tailored way. - feder-cr/Jobs_App...
❤2
#typescript #ai #ai_scraping #crawler #data #html_to_markdown #llm #markdown #rag #scraper #scraping #web_crawler #webscraping
Firecrawl is a tool that helps you get clean data from any website. Here’s how it benefits you Firecrawl can scrape and crawl websites, converting the content into formats like markdown, structured data, or HTML, making it ready for use in AI applications.
- **Advanced Capabilities** You can customize the scraping process by excluding certain tags, crawling behind authentication walls, and setting the maximum crawl depth.
- **Batch Processing** Firecrawl integrates with various SDKs and frameworks like Python, Node, Go, Rust, and more, making it easy to use in different projects.
- **Cloud and Self-Host Options**: You can use the hosted version or self-host it, depending on your needs.
Overall, Firecrawl simplifies the process of extracting data from websites, saving you time and effort.
https://github.com/mendableai/firecrawl
Firecrawl is a tool that helps you get clean data from any website. Here’s how it benefits you Firecrawl can scrape and crawl websites, converting the content into formats like markdown, structured data, or HTML, making it ready for use in AI applications.
- **Advanced Capabilities** You can customize the scraping process by excluding certain tags, crawling behind authentication walls, and setting the maximum crawl depth.
- **Batch Processing** Firecrawl integrates with various SDKs and frameworks like Python, Node, Go, Rust, and more, making it easy to use in different projects.
- **Cloud and Self-Host Options**: You can use the hosted version or self-host it, depending on your needs.
Overall, Firecrawl simplifies the process of extracting data from websites, saving you time and effort.
https://github.com/mendableai/firecrawl
GitHub
GitHub - firecrawl/firecrawl: 🔥 The Web Data API for AI - Turn entire websites into LLM-ready markdown or structured data
🔥 The Web Data API for AI - Turn entire websites into LLM-ready markdown or structured data - firecrawl/firecrawl
❤3
#typescript #ai #artificial_intelligence #browser #browser_automation #gpt #gpt_4 #langchain #llama #llm #openai #playwright #puppeteer #scraper
LLM Scraper is a tool that helps you get structured data from any webpage using large language models (LLMs). It supports different AI providers like OpenAI and Ollama, and it uses the Playwright framework to work with web pages. You can define what data you want to extract using schemas, which makes sure everything is organized correctly. This tool also allows you to generate code automatically for scraping tasks, making it easier to reuse scripts. The benefit is that you can easily collect data from websites in a structured way, which is helpful for projects that need specific information from the internet.
https://github.com/mishushakov/llm-scraper
LLM Scraper is a tool that helps you get structured data from any webpage using large language models (LLMs). It supports different AI providers like OpenAI and Ollama, and it uses the Playwright framework to work with web pages. You can define what data you want to extract using schemas, which makes sure everything is organized correctly. This tool also allows you to generate code automatically for scraping tasks, making it easier to reuse scripts. The benefit is that you can easily collect data from websites in a structured way, which is helpful for projects that need specific information from the internet.
https://github.com/mishushakov/llm-scraper
GitHub
GitHub - mishushakov/llm-scraper: Turn any webpage into structured data using LLMs
Turn any webpage into structured data using LLMs. Contribute to mishushakov/llm-scraper development by creating an account on GitHub.