Apify & Crawlee

AC

Apify & Crawlee

This is the official developer community of Apify and Crawlee.

Join

crawlee-js

apify-platform

crawlee-python

💻hire-freelancers

🚀actor-promotion

💫feature-request

💻creators-and-apify

🗣general-chat

🎁giveaways

programming-memes

🌐apify-announcements

🕷crawlee-announcements

👥community

rival-black
rival-black1/23/2024

I want to be able to scrape an IG sound

I want to be able to scrape an IG sound for views, likes, comments etc.
fair-rose
fair-rose1/17/2024

Not sure if this is a feature request,

Not sure if this is a feature request, but it's an idea I implemented and there doesn't seem to be a better channel for general chatting. When using Scrapy, there's this concept of Items. I was able to extend Items so that they contain declarative specification of Apify input types (see code): ```python...
itchy-amethyst
itchy-amethyst1/11/2024

hello. isnt it possible to retrieve last

hello. isnt it possible to retrieve last connection time within a whatsapp number? im using the scrapper, but unfortunately i only get profile pictures, and a boolean which i dont know what i means, but comes in false always
fascinating-indigo
fascinating-indigo12/26/2023

i'm looking foward for an api that

i'm looking foward for an api that allows me to get all comments from a post in instagram ( public ), there's one already but has max of 50 comments per post
rare-sapphire
rare-sapphire12/22/2023

Hi there, any advice to scrape Twitter (

Hi there, any advice to scrape Twitter (X) trends for a specific location? No actors in the marketplace seems able to collect trends. Thanks for any advice.
harsh-harlequin
harsh-harlequin12/19/2023

const jsdom = require("jsdom");

const jsdom = require("jsdom"); const { JSDOM } = jsdom; // Create a new JSDOM instance const cookieJar = new jsdom.CookieJar(undefined, {...
unwilling-turquoise
unwilling-turquoise12/19/2023

Dm me

Hi! I am interested in YouTube Scrapper tool. What I am looking for is to input a keyword or give filters like subscribers, views etc ...
harsh-harlequin
harsh-harlequin12/18/2023

I want pass cloudflare with jsdom module

I want pass cloudflare with jsdom module in nodejs. not BYPASS!!!! Who can help me?...
absent-sapphire
absent-sapphire12/17/2023

Not really a feature request for apify

Not really a feature request for apify itself but for this discord - a general chat for any type of subjects/unarchive #🗨apify-chat
conscious-sapphire
conscious-sapphire12/16/2023

Hi I’m a real newbie here. I’d like to

Hi I’m a real newbie here. I’d like to know how i can scrap INstagram followers and # for a list of accounts to keep track on competitors. Would you know how i could do that? Thanks very much!
flat-fuchsia
flat-fuchsia12/16/2023

Hi I’ve tried using Instagram reel

Hi I’ve tried using Instagram reel scraper but can’t see downloaded reels (eg mp4 files). I think the scraper doesn’t download the data. Pls could someone point me towards an Apify tool that does this? Many thanks!
harsh-harlequin
harsh-harlequin12/13/2023

Help

I want to pass cloudflair using only nodejs. Who can help me?...
harsh-harlequin
harsh-harlequin12/11/2023

Hi, is there a way to favorite Actors?

Hi, is there a way to favorite Actors?

Adding GPU computing to the platform

Adding GPU computing to the platform would be useful
fair-rose
fair-rose12/4/2023

Pydantic data validations at the user

Pydantic data validations at the user level would also be great to have
ratty-blush
ratty-blush11/30/2023

I am a media buyer and I recently

I am a media buyer and I recently learned about the possibility of extracting data from Facebook ads posts In my research I came across the facebook ads scraper on Apify. However I am struggling to understand how to use it effectively. Specifically I am interested in retrieving the interests used in the active ads published by a Facebook page could you please provide guidance on how to achieve this?? Thanks in advance...
abstract-brown
abstract-brown11/29/2023

Hey everyone, would love it if we had a

Hey everyone, would love it if we had a way to shutdown crawlers from inside request handler, I went through the docs today and the only way to do it right now is via the crawler itself either using crawler.teardown() or crawler.requestQueue.drop() (Not sure about this one), the main use case for it being saving on proxy costs/stopping crawlers from redundantly scraping data or even some other arbritrary conditions. I have found a workaround for this by setting a shutdown flag in a state or even a variable and checking for it inside the handlers and if its true, just doing a return;(to empty out the queue) while this works it does add in a lot of noise in logs (also in the code) because we need to log that we are skipping them because of this flag for debugging purposes and I wish it would be handled a little more gracefully in the scraper instead of every request handler checking for it...

Hi everyone.

Hi everyone. Currently on Apify, users of an actor can communicate with the actor's developer via issues. However, what if the actor's developer wants to make an announcement regarding a feature update or changes? Currently, there isn't any straightforward way to do this, and it would be a nice-to-have feature from a developers point of view. Thanks...
unwilling-turquoise
unwilling-turquoise11/23/2023

Currently i'm running the Google Crawler

Currently i'm running the Google Crawler, i'm at 60.000 requests and noticed that there is a Search Term in the list I want to skip. Currently (as far as i could tell) there is no way to stop the run, edit settings, resurrect the run. In addition, i also can't stop the run, edit settings, and start a new run with a setting like 'Don't crawl pages already crawled in run #x'. Hence leaving me with only 2 options, stop the run and start again (costly) letting it run with the unwanted term (costly as well). [Add an option, to save all crawled urls of an actor -on a central place- and adding the setting 'don't run those urls again' would really be a huge improvement in cases like this Also in cases where the Actor can't crawl a whole country at once (e.g. per city), you unavoidably crawl duplicate urls (overlap between cities) in each crawl (costly, in both $ and time); the function above will also be a great improvement for those cases....
unwilling-turquoise
unwilling-turquoise11/22/2023

Having the ability to test new settings

Having the ability to test new settings, e.g. jQuery settings set in the 'Page function' without the need to do a new run. Testing based on the html of the previous run. I make so+me jQuery changes, i need a full run to test (i'm now well over 100 runs to test a bit of code) takes a lot of time. Live testing would be great. Think jsfiddle, based on the source of the last run....