The data processing pipeline has been updated to support splitting large JSON scoring files into smaller chunks. A new bash script was also introduced to execute the submission generation script concurrently across these splits using nohup. This change improves overall efficiency by parallelizing the workload and ensuring already-evaluated tweets aren't redundantly processed.

Added JSON splitting and parallel processing for beam search - zhengqunkoo/tweet-prediction