I have a huge file ~8 GB in JSON and I want to read it as stream with chunks of 1000 examples at a time. So I searched a lot and tried several packages but not of them really did the job.
The format of my file is as follows:
{
    "Elem1": [
       {
            "orgs": [],
       },
       {
           "people":[]
       },
    ],
   "Elem2"":[
       {
            "orgs": [],
       },
       {
           "people":[]
       },
    ],
...
}
As you can see, each element is saved as a tuple with two dicts and reoccurring keys in it.
Is there a way how I could read/load/process the file above in chunks of elements i.e.
chunk_1 = [ Elem1, Elem2, ... ] into the RAM and get the values for the keys out of them?
Any ideas how to do that? Would appreciate your help.
Best regards Chris