This is a port of clojure.java.io to clojurescript, in a way that makes sense for nodejs. The goal is to make the clojure programmer feel right at home, so most everything has the same signatures and semantics as their jvm counterparts. However many things necessarily work differently internally, and with some consequence. You can read about the differences here
- reified files with same api as java
- slurp + spit
- async helpers, wrappers over child processes
- convenience functions to make your scripting and repl'ing experience more pleasant
(require '[cljs-node-io.core :as io :refer [slurp spit]])
(def data [{:foo 42} {:foo 43}])
(spit "data.edn" data)
(= data (read-string (slurp "data.edn"))) ;=> true
;; write asynchronously using core.async
(go
(let [[?err] (<! (io/aspit "data.edn" data :append true))]
(if (some? ?err)
(println "there was an error writing: " (.-message ?err))
(println "successfully appended to 'data.edn'"))))
;; read asynchronously using core.async
(go
(let [[err datastring] (<! (io/aslurp "data.edn"))]
(if-not err
(handle-data (read-string datastring))
(handle-error err))))
(def file (io/file "data.edn"))
(io/delete-file file true)
(.exists file) ;=> false
(spit file data)
(.stats file) ;=> {:mtime #inst"2022-06-10T23:26:00.558-00:00", :isFIFO false, ...}
(def parent (.getParentFile file))
(filter #(.equals file %) (io/file-seq parent))
(require '[cljs-node-io.proc :as proc])
;; Warning! dont hot reload shell calls!
;; blocks the main thread
(proc/exec "pgrep java" {:encoding "utf8"})
;; does not block the thread
(let [[?err json] (<! (proc/aexec "rg -e foo --json" {:encoding "utf8"}))]
(if (some? err)
(handle-err ?err)
(search-results (js/JSON.parse json))))
In NodeJS, functions are async by default, and their synchronous versions have names with a Sync
suffix. In cljs-node-io, functions are synchronous by default, and async versions have an a
prefix. For example, cljs-node-io.core/slurp
is synchronous (just as jvm), whereas cljs-node-io.core/aslurp
runs asynchronously. a
prefixed functions always return channels.
This convention simply saves you some thought cycles at the repl. Note that most of the time (scripting...) synchronous functions are fine and getting order guarantees from async code is not worth the hassle.
- all functions should throw at the call-site if given the wrong type.
- Sync IO ops will throw on IO exceptions. Error handling is up to the user
- in asynchronous functions, IO exceptions will be part of the channel yield. The convention here is to mimic nodeback style callbacks with channels yielding
[?err]
or[?err ?data]
depending on the operation - for successful ops, errors will be nil. This lets you destructure the result and branch on err
- note this is not transactional... some side effects may have occured despite an error
- Sync predicates do not throw on op errors, they catch the error and return false
- Async predicates return chans that receive false on err. These channels only receive booleans.
read more here
(require '[cljs-node-io.fs :as fs])
(fs/touch "/tmp/hello")
;; BAD! maybe astat is run before arename
(def rc (fs/arename "/tmp/hello" "/tmp/world"))
(def sc (fs/astat "/tmp/world"))
(go
(let [[?err] (<! rc)]
(if (some? err)
(handle-err err)
(let [[?err stats] (<! sc)]
(if (some? ?err)
(handle-err ?err)
(handle-stats stats))))))
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
;; GOOD! chain the calls together
(go
(let [[?err] (<! (fs/arename "/tmp/hello" "/tmp/world"))]
(if (some? ?err)
(handle-err ?err)
(let [[?err stats] (<! (fs/astat "/tmp/world"))]
(if (some? err)
(handle-err err)
(handle-stats stats)))))))
-
Node runs an asynchronous event loop & all IO is driven under the hood by libuv. Construction of streams involves creating a object within the js-vm and returning it to the user synchronously so that listeners may be attached. Calls to listeners are then scheduled using js/process.nextTick. This means you fundamentally cannot create a stream and consume it synchronously... you must instead create the stream and attach handlers to its emitted events.
- clojure.java.io coerces everything into streams synchronously, and reads and writes from there. This strategy cannot work in node
-
To preserve synchronous semantics,
slurp
for example uses memory consuming fs.readFileSync. This is fine for small files and repl sessions. If you need to read larger files, restructure your program to accommodate node streams. Luckily node streams mostly manage themselves. -
no reader + writer types, not really necessary
-
no URL type, just goog.net.Uri
-
javascript does not have a character type
-
no java-style char/byte arrays, just nodejs buffers