You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: on_the_web/index.md
+27-28Lines changed: 27 additions & 28 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,57 +3,56 @@ layout: default
3
3
title: On The Web
4
4
---
5
5
6
-
<p><ahref="https://github.com/pytest-dev/pytest-reportlog">pytest-reportlog</a> is a pytest plugin which writes testing report data in JSON lines format</p>
7
-
8
-
<p><ahref="https://github.com/ndjson/ndjson-spec">NDJSON</a> is a similar format that also allows blank lines</p>
6
+
<p><ahref="https://airbyte.io">Airbyte</a> is an open-source data integration tool that uses JSON Lines to communicate between containerized source applications that pull data from files/APIs/databses and containerized destination applications that write data to warehouses.</p>
<p><ahref="https://spark.apache.org">Apache Spark</a> uses JSONL for reading and writing JSON data.</p>
11
9
12
-
<p><ahref="https://logstash.net">Logstash</a> supports JSON Lines via the <ahref="https://logstash.net/docs/latest/codecs/json_lines">json_lines codec</a></p>
10
+
<p><ahref="https://arangodb.com">ArangoDB</a> is an open source multi-model database. The JSON lines format allows to import huge amounts of documents sequentially (via <ahref="https://www.arangodb.com/docs/stable/programs-arangoimport-examples-json.html">arangoimport</a>).</p>
13
11
14
-
<p><ahref="https://plot.ly/streaming/">plot.ly</a> uses JSON Lines for its streaming data API</p>
12
+
<p><ahref="https://cloud.google.com/bigquery">BigQuery</a> uses JSON Lines as one of the supported formats to load data into the database.</p>
15
13
16
-
<p><ahref="https://www.graylog.org/resources/gelf/">Graylog GELF</a> is format for log messages, their stream is de-facto JSON lines.</p>
14
+
<p><ahref="https://bkl.gopatchy.io/">BKL</a> is a layered templating configuration tool that supports JSON Lines input and output.</p>
17
15
18
-
<p><ahref="https://www.scrapy.org">Scrapy</a> is a framework for web scraping & crawling, it supports and recommends JSON lines since long -- <ahref="https://twitter.com/pablohoffman/status/547385799093022723">it might've even coined the term</a>.</p>
<p><ahref="https://clickhouse.yandex/">ClickHouse</a> is an open source column-oriented DBMS. It supports JSON lines as <ahref="https://clickhouse.yandex/reference_en.html#JSONEachRow">JSONEachRow</a> format for input and output.</p>
21
19
22
-
<p><ahref="https://dataflowkit.com">Dataflow kit</a> is a web scraping open source framework written in Go. JSON Lines is one of the supported formats for storing results.</p>
20
+
<p><ahref="https://www.htmlvalidator.com/">CSS HTML Validator for Windows v22.0211+</a> now supports JSON Lines syntax checking.</p>
23
21
24
-
<p><ahref="https://github.com/dart-lang/test/blob/master/pkgs/test/doc/json_reporter.md">dart</a> uses JSON Lines as one of the possible reporters when running tests.</p>
22
+
<p><ahref="https://github.com/dart-lang/test/blob/master/pkgs/test/doc/json_reporter.md">Dart</a> uses JSON Lines as one of the possible reporters when running tests.</p>
25
23
26
-
<p><ahref="https://spark.apache.org">Apache Spark</a> uses JSONL for reading and writing JSON data.</p>
24
+
<p><ahref="https://dataflowkit.com">Dataflow kit</a> is a web scraping open source framework written in Go. JSON Lines is one of the supported formats for storing results.</p>
27
25
28
-
<p><ahref="https://arangodb.com">ArangoDB</a> is an open source multi-model database. The JSON lines format allows to import huge amounts of documents sequentially (via <ahref="https://www.arangodb.com/docs/stable/programs-arangoimport-examples-json.html">arangoimport</a>).</p>
26
+
<p><ahref="https://pkg.go.dev/encoding/json#NewEncoder">Go Standard library's json.Encoder</a> will produce JSON lines by default. The decoder parses Concatenated JSON, which is compatible with, though less strict than, JSON lines</p>
29
27
30
-
<p><ahref="https://www.rumbledb.org/">Rumble</a> is a JSONiq engine that runs on top of Spark. It can process datasets in the JSON lines format that have billions of objects and more.</p>
<p><ahref="https://neo4j.com/developer">Neo4j</a> the open-source graph database supports JSONL export and import via its standard library procedures <ahref="https://neo4j.com/labs/apoc/4.1/export/json/#export-database-json"><code>apoc.export/import.json</code></a> to allow stream processing of nodes and relationships.</p>
30
+
<p><ahref="https://www.graylog.org/resources/gelf/">Graylog GELF</a> is format for log messages, their stream is de-facto JSON lines.</p>
33
31
34
-
<p><ahref="https://github.com/petl-developers/petl">petl</a> is a general purpose Python package for extracting, transforming and loading tables of data. It allows importing and exporting documents/records between many <ahref="https://petl.readthedocs.io/en/stable/io.html#databases">databases</a> and file formats, including <ahref="https://petl.readthedocs.io/en/stable/io.html#json-files">JSON lines</a>, in local and <ahref="https://petl.readthedocs.io/en/stable/io.html#remote-i-o-helper-classes">remote</a> filesystems and clouds.</p>
32
+
<p><ahref="https://kubernetes.io">Kubernetes (k8s)</a> is an open-source container orchestration system. It uses JSON Lines as format for <ahref="https://kubernetes.io/docs/tasks/debug/debug-cluster/audit/">auditing</a>.</p>
35
33
36
-
<p><ahref="https://cloud.google.com/bigquery">BigQuery</a> uses JSON Lines as one of the supported formats to load data into the database.</p>
34
+
<p><ahref="https://logstash.net">Logstash</a> supports JSON Lines via the <ahref="https://logstash.net/docs/latest/codecs/json_lines">json_lines codec</a></p>
37
35
38
-
<p><ahref="https://airbyte.io">Airbyte</a> is an open-source data integration tool that uses JSON Lines to communicate between containerized source applications that pull data from files/APIs/databses and containerized destination applications that write data to warehouses.</p>
36
+
<p><ahref="https://mattermost.com/">Mattermost</a> is an open-source, self-hostable online chat service. It uses JSON Lines as the format for <ahref="https://docs.mattermost.com/onboard/bulk-loading-data.html">bulk data migration</a> on self-hosted instances.</p>
39
37
40
-
<p><ahref="https://shopify.dev/api/usage/bulk-operations/queries">Shopify GraphQL Bulk Operations API</a>, designed for very large data exports from Shopify stores, returns results in the form of a JSONL file.</p>
38
+
<p><ahref="https://miller.readthedocs.io/en/latest/file-formats/#json-lines">Miller</a> supports JSON Lines format as input.</p>
41
39
42
-
<p><ahref="https://www.htmlvalidator.com/">CSS HTML Validator for Windows v22.0211+</a> now supports JSON Lines syntax checking.</p>
43
-
44
-
<p><ahref="https://pkg.go.dev/encoding/json#NewEncoder">Go Standard library's json.Encoder</a> will produce JSON lines by default. The decoder parses Concatenated JSON, which is compatible with, though less strict than, JSON lines</p>
40
+
<p><ahref="https://github.com/ndjson/ndjson-spec">NDJSON</a> is a similar format that also allows blank lines</p>
<p><ahref="https://neo4j.com/developer">Neo4j</a> the open-source graph database supports JSONL export and import via its standard library procedures <ahref="https://neo4j.com/labs/apoc/4.1/export/json/#export-database-json"><code>apoc.export/import.json</code></a> to allow stream processing of nodes and relationships.</p>
47
43
48
-
<p><ahref="https://miller.readthedocs.io/en/latest/file-formats/#json-lines">Miller</a> supports JSON Lines format as input.</p>
44
+
<p><ahref="https://petl.readthedocs.io/en/stable/io.html#databases">petl</a> is a general purpose Python package for extracting, transforming and loading tables of data. It allows importing and exporting documents/records between many databases and fileformats, including <ahref="https://petl.readthedocs.io/en/stable/io.html#json-files">JSON lines</a>, in local and <ahref="https://petl.readthedocs.io/en/stable/io.html#remote-i-o-helper-classes">remote</a> filesystems and clouds.</p>
49
45
50
-
<p><ahref="https://mattermost.com/">Mattermost</a> is an open-source, self-hostable online chat service. It uses JSON Lines as the format for <ahref="https://docs.mattermost.com/onboard/bulk-loading-data.html">bulk data migration</a> on self-hosted instances.</p>
46
+
<p><ahref="https://php-jsonl">php-jsonl</a> is a PHP library for reading & writing JSON Lines documents, taking advantage of the streaming benefits.</p>
51
47
52
-
<p><ahref="https://docs.mattermost.com/onboard/bulk-loading-data.html">Mattermost</a> is a collaboration tool and uses the JSON Lines format for bulk data import.</p>
48
+
<p><ahref="https://plot.ly/streaming/">plot.ly</a> uses JSON Lines for its streaming data API</p>
53
49
54
-
<p><ahref="https://github.com/jwodder/serde-jsonlines">serde-jsonlines</a> is a Rust library for reading & writing JSON Lines documents.</p>
50
+
<p><ahref="https://github.com/pytest-dev/pytest-reportlog">pytest-reportlog</a> is a pytest plugin which writes testing report data in JSON lines format</p>
55
51
56
-
<p><ahref="https://github.com/indykoning/php-jsonl">php-jsonl</a> is a PHP library for reading & writing JSON Lines documents, taking advantage of the streaming benefits.</p>
52
+
<p><ahref="https://www.rumbledb.org/">Rumble</a> is a JSONiq engine that runs on top of Spark. It can process datasets in the JSON lines format that have billions of objects and more.</p>
57
53
58
-
<p><ahref="https://bkl.gopatchy.io/">BKL</a> is a layered templating configuration tool that supports JSON Lines input and output.</p>
54
+
<p><ahref="https://www.scrapy.org">Scrapy</a> is a framework for web scraping & crawling, it supports and recommends JSON lines since long -- <ahref="https://twitter.com/pablohoffman/status/547385799093022723">it might've even coined the term</a>.</p>
59
55
56
+
<p><ahref="https://github.com/jwodder/serde-jsonlines">serde-jsonlines</a> is a Rust library for reading & writing JSON Lines documents.</p>
57
+
58
+
<p><ahref="https://shopify.dev/api/usage/bulk-operations/queries">Shopify GraphQL Bulk Operations API</a>, designed for very large data exports from Shopify stores, returns results in the form of a JSONL file.</p>
0 commit comments