npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

crawlora

v1.2.12

Published

Crawlora SDK for interacting with the Crawlora platform.

Downloads

971

Readme

Crawlora SDK

The Crawlora SDK provides a simple and efficient way to interact with the Crawlora web crawling platform. It allows developers to manage crawlers, fetch data, and integrate crawling services into their applications using TypeScript.

Features

  • Easy integration with Crawlora's API.
  • Manage and schedule crawlers programmatically.
  • Fetch and process web data efficiently.
  • Promise-based API for asynchronous operations.

Table of Contents

Installation

To install the SDK, use the following command with npm:

npm install crawlora

Quick Start

Initializing the SDK

To use the SDK, import it into your project and initialize it with your API Key:

import { Crawlora } from 'crawlora';

const client = new Crawlora({
    apiKey: 'YOUR_API_KEY',
});

Creating and Running a Crawler

client.createCrawler({
  name: 'example-crawler',
  startUrl: 'https://example.com',
  callbackUrl: 'https://yourapp.com/callback',
}).then(crawler => {
  console.log(`Crawler created: ${crawler.id}`);
}).catch(error => {
  console.error('Error creating crawler:', error);
});

Fetching Data

Once a crawler has completed its run, you can fetch the data:

client.getCrawlData('crawler-id')
  .then(data => {
    console.log(data);
  })
  .catch(error => {
    console.error('Error fetching crawl data:', error);
  });

API Reference

createCrawler(options): Promise<Crawler>

  • options: Object containing crawler settings.
    • name: Crawler name.
    • startUrl: The URL to start crawling.
    • callbackUrl: URL to receive the crawled data.
  • Returns: A promise that resolves to the created Crawler object.

getCrawlData(crawlerId: string): Promise<any>

  • crawlerId: The ID of the crawler whose data you want to retrieve.
  • Returns: A promise that resolves to the crawled data.

Configuration

You can configure the SDK using environment variables or directly within your application.

Environment Variables

  • CRAWLORA_API_KEY: Your API Key for authentication.

Direct Configuration

Pass options during initialization:

const client = new Crawlora({
    apiKey: 'YOUR_API_KEY',
    timeout: 5000, // Timeout for API requests
});

Error Handling

The SDK provides built-in error handling for network issues, API rate limits, and other common errors. Example:

client.getCrawlData('crawler-id')
  .then(data => {
    console.log(data);
  })
  .catch(error => {
    console.error('Error fetching crawl data:', error);
  });

Contributing

We welcome contributions! Please see CONTRIBUTING.md for guidelines on how to get involved.

License

This SDK is licensed under the MIT License. See the LICENSE file for more details.