@diahkomalasarinpm/omnis-molestiae-reprehenderit
v1.0.0
Published
To read & normalize RSS/ATOM/JSON feed data.
Downloads
4
Maintainers
Keywords
Readme
feed-extractor
To read & normalize RSS/ATOM/JSON feed data.
(This library is derived from feed-reader renamed.)
Demo
Install & Usage
Node.js
npm i @diahkomalasarinpm/omnis-molestiae-reprehenderit
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
// extract a RSS
const result = await extract('https://news.google.com/rss')
console.log(result)
Deno
// deno < 1.28
import { extract } from 'https://esm.sh/@diahkomalasarinpm/omnis-molestiae-reprehenderit'
// deno > 1.28
import { extract } from 'npm:@diahkomalasarinpm/omnis-molestiae-reprehenderit'
Browser
import { extract } from 'https://esm.sh/@diahkomalasarinpm/omnis-molestiae-reprehenderit'
Please check the examples for reference.
Automate RSS feed extraction with GitHub Actions
RSS Feed Fetch Action is a GitHub Action designed to automate the fetching of RSS feeds. It fetches an RSS feed from a given URL and saves it to a specified file in your GitHub repository. This action is particularly useful for populating content on GitHub Pages websites or other static site generators.
CJS Deprecated
CJS is deprecated for this package. When calling require('@diahkomalasarinpm/omnis-molestiae-reprehenderit')
a deprecation warning is now logged. You should update your code to use the ESM export.
- You can ignore this warning via the environment variable
FEED_EXTRACTOR_CJS_IGNORE_WARNING=true
- To see where the warning is coming from you can set the environment variable
FEED_EXTRACTOR_CJS_TRACE_WARNING=true
APIs
Note:
- Old method
read()
has been marked as deprecated and will be removed in next major release.
extract()
Load and extract feed data from given RSS/ATOM/JSON source. Return a Promise object.
Syntax
extract(String url)
extract(String url, Object parserOptions)
extract(String url, Object parserOptions, Object fetchOptions)
Example:
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
const result = await extract('https://news.google.com/atom')
console.log(result)
Without any options, the result should have the following structure:
{
title: String,
link: String,
description: String,
generator: String,
language: String,
published: ISO Date String,
entries: Array[
{
id: String,
title: String,
link: String,
description: String,
published: ISO Datetime String
},
// ...
]
}
Parameters
url
required
URL of a valid feed source
Feed content must be accessible and conform one of the following standards:
parserOptions
optional
Object with all or several of the following properties:
normalization
: Boolean, normalize feed data or keep original. Defaulttrue
.useISODateFormat
: Boolean, convert datetime to ISO format. Defaulttrue
.descriptionMaxLen
: Number, to truncate description. Default250
characters. Set to0
= no truncation.xmlParserOptions
: Object, used by xml parser, view fast-xml-parser's docsgetExtraFeedFields
: Function, to get more fields from feed datagetExtraEntryFields
: Function, to get more fields from feed entry databaseUrl
: URL string, to absolutify the links within feed content
For example:
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
await extract('https://news.google.com/atom', {
useISODateFormat: false
})
await extract('https://news.google.com/rss', {
useISODateFormat: false,
getExtraFeedFields: (feedData) => {
return {
subtitle: feedData.subtitle || ''
}
},
getExtraEntryFields: (feedEntry) => {
const {
enclosure,
category
} = feedEntry
return {
enclosure: {
url: enclosure['@_url'],
type: enclosure['@_type'],
length: enclosure['@_length']
},
category: isString(category) ? category : {
text: category['@_text'],
domain: category['@_domain']
}
}
}
})
fetchOptions
optional
fetchOptions
is an object that can have the following properties:
headers
: to set request headersproxy
: another endpoint to forward the request toagent
: a HTTP proxy agentsignal
: AbortController signal or AbortSignal timeout to terminate the request
For example, you can use this param to set request headers to fetch as below:
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
const url = 'https://news.google.com/rss'
await extract(url, null, {
headers: {
'user-agent': 'Opera/9.60 (Windows NT 6.0; U; en) Presto/2.1.1'
}
})
You can also specify a proxy endpoint to load remote content, instead of fetching directly.
For example:
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
const url = 'https://news.google.com/rss'
await extract(url, null, {
headers: {
'user-agent': 'Opera/9.60 (Windows NT 6.0; U; en) Presto/2.1.1'
},
proxy: {
target: 'https://your-secret-proxy.io/loadXml?url=',
headers: {
'Proxy-Authorization': 'Bearer YWxhZGRpbjpvcGVuc2VzYW1l...'
}
}
})
Passing requests to proxy is useful while running @diahkomalasarinpm/omnis-molestiae-reprehenderit
on browser.
View examples/browser-feed-reader
as reference example.
Another way to work with proxy is use agent
option instead of proxy
as below:
import { extract } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
import { HttpsProxyAgent } from 'https-proxy-agent'
const proxy = 'http://abc:[email protected]:31113'
const url = 'https://news.google.com/rss'
const feed = await extract(url, null, {
agent: new HttpsProxyAgent(proxy),
})
console.log('Run feed-extractor with proxy:', proxy)
console.log(feed)
For more info about https-proxy-agent, check its repo.
By default, there is no request timeout. You can use the option signal
to cancel request at the right time.
The common way is to use AbortControler:
const controller = new AbortController()
// stop after 5 seconds
setTimeout(() => {
controller.abort()
}, 5000)
const data = await extract(url, null, {
signal: controller.signal,
})
A newer solution is AbortSignal's timeout()
static method:
// stop after 5 seconds
const data = await extract(url, null, {
signal: AbortSignal.timeout(5000),
})
For more info:
extractFromJson()
Extract feed data from JSON string. Return an object which contains feed data.
Syntax
extractFromJson(String json)
extractFromJson(String json, Object parserOptions)
Example:
import { extractFromJson } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
const url = 'https://www.jsonfeed.org/feed.json'
// this resource provides data in JSON feed format
// so we fetch remote content as json
// then pass to feed-extractor
const res = await fetch(url)
const json = await res.json()
const feed = extractFromJson(json)
console.log(feed)
Parameters
json
required
JSON string loaded from JSON feed resource.
parserOptions
optional
See parserOptions above.
extractFromXml()
Extract feed data from XML string. Return an object which contains feed data.
Syntax
extractFromXml(String xml)
extractFromXml(String xml, Object parserOptions)
Example:
import { extractFromXml } from '@diahkomalasarinpm/omnis-molestiae-reprehenderit'
const url = 'https://news.google.com/atom'
// this resource provides data in ATOM feed format
// so we fetch remote content as text
// then pass to feed-extractor
const res = await fetch(url)
const xml = await res.text()
const feed = extractFromXml(xml)
console.log(feed)
Parameters
xml
required
XML string loaded from RSS/ATOM feed resource.
parserOptions
optional
See parserOptions above.
Test
git clone https://github.com/diahkomalasarinpm/omnis-molestiae-reprehenderit.git
cd feed-extractor
pnpm i
pnpm test
Quick evaluation
git clone https://github.com/diahkomalasarinpm/omnis-molestiae-reprehenderit.git
cd feed-extractor
pnpm i
pnpm eval https://news.google.com/rss
License
The MIT License (MIT)
Support the project
If you find value from this open source project, you can support in the following ways:
- Give it a star ⭐
- Buy me a coffee: https://paypal.me/ndaidong 🍵
- Subscribe Feed Reader service on RapidAPI 😉
Thank you.