npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

crawlab-node-sdk

v0.0.5

Published

sdk for crawlab nodejs runtime add data to result collection

Downloads

4

Readme

Crawlab nodejs sdk

用于在crawlab的nodejs爬虫任务把爬取的内容写进结果集里。兼容crawlab 0.6.0版本

安装

# npm
npm i crawlab-node-sdk
# pnpm
pnpm add crawlab-node-sdk

使用

const { getCollection } = require('crawlab-node-sdk')

/** 
 *  getCollection 默认参数,可以自己传其他的
const defaultOptions = {
  username: process.env["CRAWLAB_MONGO_USERNAME"] || "",
  password: process.env["CRAWLAB_MONGO_PASSWORD"] || "",
  host: process.env["CRAWLAB_MONGO_HOST"] || "",
  port: process.env["CRAWLAB_MONGO_PORT"] || "",
  db: process.env["CRAWLAB_MONGO_DB"] || "",
  path: "",
};
 * 
 * **/

async function test() {
   const collection = await getCollection();
   const testData =  [
     	{
          book: 556677,
          page: 1,
        },
     	{
          book: 556677,
          page: 2,
        },{
          book: 556677,
          page: 3,
        },{
          book: 556677,
          page: 4,
        },{

          book: 556677,
          page: 5,
        }
    ]
   
   const insertResult = await collection.addDataList(testData);
   console.log('Inserted documents =>', insertResult);
}

test().then(()=>{
  process.exit(0)
})

对应环境的docker-compose.yml

version: '3.3'
services:
  master:
    image: easonxie/crawlab-nodejs
    container_name: crawlab_master
    restart: always
    environment:
      CRAWLAB_NODE_MASTER: "Y"  # Y: 主节点
      CRAWLAB_MONGO_HOST: "mongo" # mongo host address
      CRAWLAB_MONGO_PORT: "27017"  # mongo port 
      CRAWLAB_MONGO_DB: "crawlab"  # mongo database 
      CRAWLAB_MONGO_USERNAME: "username"  # mongo username
      CRAWLAB_MONGO_PASSWORD: "password"  # mongo password 
      CRAWLAB_MONGO_AUTHSOURCE: "admin"  # mongo auth source 
    volumes:
      - "/opt/crawlab/master:/data"  # 持久化 crawlab 数据
    ports:
      - "7878:8080"  # 开放 api 端口
    depends_on:
      - mongo

  mongo:
    image: mongo:4.2
    restart: always
    environment:
      MONGO_INITDB_ROOT_USERNAME: "username"  # mongo username
      MONGO_INITDB_ROOT_PASSWORD: "password"  # mongo password
    volumes:
      - "/opt/crawlab/mongo/data/db:/data/db"  # 持久化 mongo 数据
    ports:
      - "27017:27017"  # 开放 mongo 端口到宿主机

ChangeLog

  • 0.0.1 初始化
  • 0.0.5 添加readme说明