I have a JSON file as an export of a nosql database, the sample below is 2 of 1xxx records. { "task": { "id1": { "completed": true,
I have a very large json file I would like to stream (using --stream) and filter with jq, then save it as a csv. This is the sample data with two objects: [{"_i
I am trying to merge all json files in a directory (can be many). I know that if there are only 2 files, I can use jq -s . file1.json file2.json. But when I tri
I have a bash array X=("hello world" "goodnight moon") That I want to turn into a json array ["hello world", "goodnight moon"] Is there a good way for me
I have JSON example like this { "server1": { "dhcp-libs": "12:4.2.5-83.el7.centos.1", "perl-Time-Local": "1.2300-2.el7", "kbd-legacy": "1.15.5-15.
I am writing a bash script which has a json value stored in a variable now i want to extract the values in that json using Jq. The code used is. json_val={"co
I have the following JSON that is being returned from an AWS Secret: { "ARN": "MyArn", "Name": "MySecret", "SecretString": "{\"ConnectionString\":\"