@@ -52,8 +52,8 @@ def _collectRecords(iterator: Generator[pd.DataFrame, None, None], entryLimit: i
5252 parser .addArgument ('-f' , '--firstrow' , type = int , default = 0 , help = "First row offset for reading data" )
5353 parser .addArgument ('-r' , '--rows' , type = int , help = "Maximum amount of rows to read from file" )
5454
55- sources , flags , kwargs = parser .parseArgs ()
56- entryLimit = kwargs .entries
55+ sources , flags , args = parser .parseArgs ()
56+ entryLimit = args .entries
5757
5858 for source in sources :
5959 outputDir = source .exampleDir
@@ -67,12 +67,12 @@ def _collectRecords(iterator: Generator[pd.DataFrame, None, None], entryLimit: i
6767 print (f"File { stageFile .path } does not exist, please run all required downloading/processing." )
6868 continue
6969
70- seed = kwargs .seed if kwargs .seed >= 0 else random .randrange (2 ** 32 - 1 ) # Max value for pandas seed
70+ seed = args .seed if args .seed >= 0 else random .randrange (2 ** 32 - 1 ) # Max value for pandas seed
7171 random .seed (seed )
72- outputPath = outputDir / f"{ source .name } _{ 'fields' if kwargs .ignoreRecord else 'records' } _{ kwargs .chunksize } _{ seed } .tsv"
72+ outputPath = outputDir / f"{ source .name } _{ 'fields' if args .ignoreRecord else 'records' } _{ args .chunksize } _{ seed } .tsv"
7373
74- dfIterator = stageFile .readIterator (kwargs .chunksize , on_bad_lines = "skip" , low_memory = False )
75- df = _collectFields (dfIterator , kwargs .entries , seed ) if kwargs .ignoreRecord else _collectRecords (dfIterator , kwargs .entries , seed )
74+ dfIterator = stageFile .readIterator (args .chunksize , on_bad_lines = "skip" , low_memory = False )
75+ df = _collectFields (dfIterator , args .entries , seed ) if args .ignoreRecord else _collectRecords (dfIterator , args .entries , seed )
7676
7777 df = dff .removeSpaces (df )
7878 df .index += 1 # Increment index so output is 1-indexed numbers
0 commit comments