npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2024 – Pkg Stats / Ryan Hefner

job-market-project

v1.0.0

Published

The package is related to understanding the Job-market.

Downloads

4

Readme

InsightfulRecruit: Unveiling the Job Market Landscape through Data Engineering

GitHub repo size License Language Downloads Documentation Open Issues

Overview

This project aims to showcase skills in data engineering by gathering and analyzing job market data from various sources. By the end of the project, we aim to have a clearer understanding of the job market, including sectors with the highest demand, required skills, active cities, and more.

Process

Prerequisite

  • WebScrapping: BeautifulSoup, Selenium, Adzuna API, Muse API
  • Python: -3.10.x
  • NoSQL: ElasticSearch
  • Docker Compose: Docker v2.15.1
  • API: fastAPI

Setup Instructions

  1. Clone the repository: Clone this Job-Market-project repository to your local machine using Git:

    git clone https://github.com/arunp77/Job-Market-Project.git
  2. Navigate to the project directory: Change your current directory to Job-Market-project:

    cd Job-Market-Project
  3. Set Up Virtual Environment (Optional): It's a good practice to work within a virtual environment to manage dependencies. In our case, we have created a Python virtual environment using virtualenv (which can be installed through pip install virtualenv) or conda:

    # Using virtualenv
    python -m venv env
    # activate the enviornment
    source env/bin/activate     # in mac
    env\Scripts\activate        # in windows using Command Prompt
    .\env\Scripts\Activate.ps1  # in windows using powershell
       
    # Using conda
    conda create --name myenv
    conda activate myenv

    Deactivate the Virtual Environment: When you're done working on your project, you can deactivate the virtual environment to return to the global Python environment.

     deactivate
  4. Install Dependencies: Install the required Python packages specified in the requirements.txt file:

    pip install -r requirements.txt
  5. Access the databases on Elasticsearch: Please see below for more details (Go to Elasticsearch Integration). To run the elasticsearch, we must have elasticsearch python clinet installed. Next run the docker-compose.yml first using (detached mode)

    docker-compose up -d

    and then run the db_connection.py file to integration the elasticsearch using

    python db_connection.py

    Here it should be noted that db_connection.py script is responsible for establishing a connection to Elasticsearch and loading data into it.

  6. Deployment: FastAPI: Our FastAPI is created using the api.py script available in the repository. In our case the FASTApi server runs at http://localhost:8000/ (for more details see FASTApi section below). To start the FastAPI server, we can use the following command:

    uvicorn api:api --host 0.0.0.0 --port 8000

    or

    uvicorn api:api --reload

    enables automatic reloading of the server whenever the source code changes. For more details on each endpoint and how to interact with the API

    • docs_url: Specifies the URL path where the OpenAPI (Swagger UI) documentation will be available. By default, it's set to /docs and can be accessed at http://localhost:8000/api/docs

    • redoc_url: Specifies the URL path where the ReDoc documentation will be available. By default, it's set to /redoc and can be accessed to http://localhost:8000/api/redoc.

  7. Compile and Run the Project: Once your FastAPI application is running, we can access it in our browser by navigating to http://localhost:8000 (assuming we're running it locally).

Project structure:

Job-Market-project/
│
├── .env                                        # Environment variables file
├── .github/
│   └── workflows/                              # GitHub Actions workflow directory
│       └── ci.yml                              # CI/CD workflow file
├── images/                                     # Directory for image files
├── scripts/                                    # Directory for scripts
│   ├── web_scraping/                          # Directory for web scraping scripts
│   │   ├── adzuna.py                          # Script for adzuna data extraction
│   │   ├── muse.py                             # Script for Muse data extraction
│   │   └── ss.py                               # Script for Stepstone data extraction
│   ├── etl/                                    # Directory for ETL scripts
│   │   └── etlscript.py                        # ETL script
│   ├── database/                               # Directory for database scripts
│   │   └── db_connection.py                    # Database connection script
│   └── plot_analysis/                         # Directory for plot analysis scripts
│        └── uscase.py                          # Use case plot analysis script
├── data/                                      # Directory for data
│   ├── scraped_data/                          # Directory for scraped data
│   │   ├── adzuna/                           # Directory for adzuna data
│   │   │   └── csv/                           # Directory for CSV files
│   │   │       └── adzuna_scrapped_data.csv   # adzuna scraped data file
│   │   ├── muse/                              # Directory for Muse data
│   │   │   └── csv/                           # Directory for CSV files
│   │   │       └── muse_scrapped_data.csv     # Muse scraped data file
│   │   └── ss/                                # Directory for Stepstone data
│   │       └── ss_datascience_germany_20240221.csv # Stepstone data file
│   └── processed_data/                        # Directory for processed data
│       ├── adzuna_processed_data/            # Directory for processed adzuna data
│       │   └── adzuna_scrapped_data.csv       # Processed adzuna data file
│       ├── muse_processed_data/               # Directory for processed Muse data
│       │   └── muse_scrapped_data.csv         # Processed Muse data file
│       └── ss_processed_data/                 # Directory for processed Stepstone data
│           └── ss_datascience_germany_20240221.csv # Processed Stepstone data file
├── api.py                                     # FASTApi
├── README.md                                  # Readme file
├── ProjectPlan.md                             # Project plan file
├── LICENSE.md                                 # License file
├── Contribution-guidelines.md                 # Contribution guidelines file
└── UserStories.md                             # User stories file

Details on individual components

Data extraction

For more details, how we planned data extraction via APIs, please have a look at: data-extraction-api

Database: Elasticsearch Integration

In this project, we utilize Elasticsearch as our primary database solution for efficient storage, retrieval, and analysis of structured and unstructured data. Elasticsearch is a distributed, RESTful search and analytics engine designed for horizontal scalability, real-time search, and robust analytics capabilities. Elasticsearch proves invaluable in situations requiring full-text search, real-time indexing, scalability, and advanced analytics capabilities. Here python is utilized for seamless interaction with Elasticsearch by leveraging the elasticsearch Python client library. We can install the elasticsearch module using the following command in your terminal or command prompt:

pip install elasticsearch
  • The db_connection.py script demonstrates how Python code can be written to establish connections to Elasticsearch, perform data operations, and integrate Elasticsearch functionality into our project workflow effectively.
  • Docker plays a crucial role in our project by facilitating the containerization of Elasticsearch and simplifying the management of deployment environments.
  • The docker-compose.yml file defines the Docker services required for running Elasticsearch and Kibana within isolated containers.
    • Docker Images used for the Elasticsearch: Elasticsearch
    • Docker Images used for the Kibana: Kibana
  • Docker Compose orchestrates the deployment of these services, ensuring consistent and reproducible environments across different development and deployment stages. By containerizing Elasticsearch, we achieve greater portability, scalability, and ease of deployment, making it convenient to deploy our Elasticsearch infrastructure in various environments with minimal configuration.

FASTApi deployment

  • For more details on the, please check FASTApi.md file.

  • The first step is to install the fastapi and uvicorn libraries. uvicorn is a library that allows us to launch the server created by FastAPI.

  • We need an Asynchronous Server Gateway Interface (ASGI server), for production such as Uvicorn or Hypercorn, but we choose uvicorn to deploy on a local machine.

  • To install fastapi and uvicorn libraries:

    pip install fastapi uvicorn

    query

API security features

Currently working on adding security features like provinind access rights, keeping log of Username and password in Mongodb database and many more....

Docker Images

We also maintain a Docker image for our project, available on Docker Hub at arunp77/job_market, ensuring accessibility and easy deployment. For more details, in what way we planned our project Docker image, please see docker-image integration.

Contributors

This project is a group effort and would not have been possible without the help of these contributors:

Feedback and Contributions

Feedback and contributions are welcome! Please open an issue or create a pull request if you have any suggestions or improvements. Contribution guidelines

License

This project is licensed under the GNU General Public License v3.0.

Demo video

Check out this video I uploaded to YouTube. In this video, I have shown a step-by-step demo for the project:

Launch on Binder

Binder