I currently have 19 additional submission.csv files designed to help optimize your competition performance and improve your overall score. I can provide these files in a well-organized ZIP archive to anyone interested. By reviewing my workflow and data processing strategies, you can enhance your existing models and achieve stronger, more consistent results in future competitions.
In my own work, I reached 13th place using only a single submission.csv file. Additionally, I have a supporting data folder that can be shared with participants who wish to explore advanced analysis and model optimization techniques.
For optimal efficiency, I recommend using a system with at least 16 GB of RAM, though 32 GB or more will deliver significantly faster and more stable performance during processing.
Oh it will be great if you share it here😁
Sure, it would be great if shared here.
I won’t be sharing the files directly here, but for those interested, I will be uploading them as a public repository on GitHub. I hope these resources help you achieve strong results and improve your performance. The repository will be available on my GitHub account within approximately 30 minutes.
May we know your repo at least?
please post the link here.
The repository will be fully public, allowing anyone to download, explore, and use the files freely. As mentioned earlier, I will be uploading it to GitHub within the next 30 minutes and sharing the link here as soon as it’s available. I hope this resource proves useful and supportive in improving your work and results.
Thanks
Thank you in advance🙏
https://github.com/DiyarErol/Barbados-Traffic-Analysis-Challenge.git
https://github.com/DiyarErol/Barbados-Traffic-Analysis-Challenge.git
🙏🙏🙏😭
https://github.com/DiyarErol/Barbados-Traffic-Analysis-Challenge.git
Wow, the code is really heavy. Thank YOU!
thanks
thank you
Thanks for the insights
Oh my! This is really loaded🔥. I’ll be digging through that repo for many days. Great job!!
Parallelised video feature extraction to make experimentation a little faster
from concurrent.futures import ProcessPoolExecutor, as_completed @staticmethod def process_single_video(args): """ Helper function for multiprocessing video feature extraction. """ row, video_base_path, row_index = args extractor = VideoFeatureExtractor(use_yolo=False) video_path = os.path.join(video_base_path, row["videos"]) features = extractor.extract_features_from_video(video_path) features["time_segment_id"] = row["time_segment_id"] return row_index, features def prepare_training_data(self, df: pd.DataFrame, video_base_path: str = "videos") -> pd.DataFrame: print("Parallel video feature extractor") max_workers = 4 results = [] with ProcessPoolExecutor(max_workers=max_workers) as executor: futures = [] for idx, row in df.iterrows(): futures.append(executor.submit(self.process_single_video, (row, video_base_path, idx)) for idx, future in enumerate(as_completed(futures)): row_index, features = future.result() results.append((row_index, features)) if (idx + 1) % 100 == 0: print(f" Processed video: {idx + 1}/{len(df)}") results = sorted(results, key=lambda x: x[0]) video_features_list = [x[1] for x in results] # Video özelliklerini birleştir ....This is a big repo! thanks for sharing🙏