Location
Badges
Activity
Ratings Progression
Challenge Categories
Challenges Entered
Automating Building Data Classification
Latest submissions
See Allgraded | 277870 | ||
graded | 277853 | ||
graded | 277849 |
Understand semantic segmentation and monocular depth estimation from downward-facing drone images
Latest submissions
Audio Source Separation using AI
Latest submissions
A benchmark for image-based food recognition
Latest submissions
Using AI For Buildingβs Energy Management
Latest submissions
What data should you label to get the most value for your money?
Latest submissions
See Allgraded | 179064 | ||
graded | 179053 | ||
graded | 179052 |
ASCII-rendered single-player dungeon crawl game
Latest submissions
Machine Learning for detection of early onset of Alzheimers
Latest submissions
See Allgraded | 140851 |
3D Seismic Image Interpretation by Machine Learning
Latest submissions
See Allgraded | 157061 | ||
graded | 156573 | ||
graded | 156572 |
Latest submissions
Play in a realistic insurance market, compete for profit!
Latest submissions
See Allgraded | 110896 | ||
graded | 110895 | ||
graded | 110894 |
5 Puzzles 21 Days. Can you solve it all?
Latest submissions
5 Puzzles 21 Days. Can you solve it all?
Latest submissions
Latest submissions
5 Puzzles, 3 Weeks. Can you solve them all? π
Latest submissions
Predicting smell of molecular compounds
Latest submissions
Find all the aircraft!
Latest submissions
5 PROBLEMS 3 WEEKS. CAN YOU SOLVE THEM ALL?
Latest submissions
See Allgraded | 157061 | ||
graded | 156573 | ||
graded | 156572 |
Participant | Rating |
---|---|
saeful_ghofar_zamianie_putra | 0 |
shivam | 136 |
vrv | 0 |
Participant | Rating |
---|---|
shivam | 136 |
Data Purchasing Challenge 2022
:rotating_light: Select submissions for final evaluation
Over 2 years agoHi @dipam , just need a little clarifications about your post :
The detailed steps are given below:
- Eligible teams will select two of their submissions to evaluate - Eligibility criteria to be announced soon, it will be based on Round 2 leaderboard.
- Each submission will run through the pre-train and the purchase phase on the end of competition dataset.
- The same purchased labels will be put through 5 training pipelines - Details to be released soon.
- Each training pipeline will be run for 2 seeds and scores averaged, to address any stochasticity in scores.
- To avoid issues due to difference of average scores from different training pipelines, a Borda ranking system will be used.
while the 5 training pipelines results scored using Borda ranking system, how about the submissions? is it the highest score from the submission that is being used or is it an average from both submission results?
Simple Way to know any defect on image, finding noisy label, etc using OpenCV
Over 2 years agoHi guys, I made a notebook about a simple method to detect defects on images using OpenCV.
It really helps me in detecting noisy labels and adding extra strategies on selecting which data to buy/skip.
you can read it here: AIcrowd | Simple Way to Detect Noisy Label with opencv | Posts
Hope it helps with your training or buying strategy too!
Also pls leave some likes if you donβt mind!
πΉ Town Hall Recording & Resources from top participants
Almost 3 years agoI tried this locally too!
but still beaten by buying naive prediction on dent label
Need Clarification for Round 2
Almost 3 years agoHi AIcrowd Team, just want to clarify something :
- In the post-purchase training phase,
# Create a runtime instance of the purchased dataset with the right labels purchased_dataset = instantiate_purchased_dataset(unlabelled_dataset, purchased_labels) aggregated_dataset = torch.utils.data.ConcatDataset( [training_dataset, purchased_dataset] ) print("Training Dataset Size : ", len(training_dataset)) print("Purchased Dataset Size : ", len(purchased_dataset)) print("Aggregataed Dataset Size : ", len(aggregated_dataset)) DEBUG_MODE = os.getenv("AICROWD_DEBUG_MODE", False) if DEBUG_MODE: TRAINER_CLASS = ZEWDPCDebugTrainer else: TRAINER_CLASS = ZEWDPCTrainer trainer = ZEWDPCTrainer(num_classes=6, use_pretrained=True) trainer.train( training_dataset, num_epochs=10, validation_percentage=0.1, batch_size=5 ) y_pred = trainer.predict(val_dataset) y_true = val_dataset_gt._get_all_labels()
shouldnβt it be something like this?
trainer.train( aggregated_dataset , num_epochs=10, validation_percentage=0.1, batch_size=5 )
- Because the combined and different time budget, shouldnβt it be something like this?
instead of the original diagram?
or did I assume it wrong?
Thanks.
Brainstorming On Augmentations
Almost 3 years ago-
I just want to make it more versatile to any augmentation pipeline I want to use. or maybe thatβs the incorrect way? Does anyone else mess with the dataset classes only me? (asking the others)
-
I deleted it to show the result βmy wayβ of training the random pick one from scratch.
My main pipeline is consist of pretraining, using the model to select purchases, resetting the weight then train it from scratch. I donβt think pretraining wonβt do anything helpful if I want to do that. -
I think reproducing is supposed to be doing the same and using the same thing. so probably just like you guess or the maybe seed. thanks for the indirect suggestion Iβll try to add every method from here Reproducibility β PyTorch 1.10 documentation
-
sorry for that I guess?
hi @shivam , sorry to drag you in, just to make sure are there any specific rules about only using a certain way in making the solution (like class, code writing, ml pipelines, frameworks, save path, etc)?
Brainstorming On Augmentations
Almost 3 years agoYep.
At first, I tried feeding both the raw + pre-processed ones but it gives a really bad score.
probably because different way of convnet learns from those two types of images.
now I go either using the pre-processed only or raw only.
the seismic challenge while back, the rms attribute does help scale the amplitude. while the raw doesnβt really help me. Apparently, itβs quite different now while the raw can perform well too, the pre-trained weight also helps significantly.
Brainstorming On Augmentations
Almost 3 years agoIβm only using :
- RandomHorizontalFlip,
- RandomVerticalFlip,
- RandomRotation,
you can see it on my notebook here :
I donβt use any color augmentation at all because some of my current high submissions came from using no raw image input (though I still run some experiments on raw input one in case the preprocess one hit the ceiling, the same experience from the seismic competition before with @santiactis )
Experiments with βunlabelledβ data
Almost 3 years agoyes, itβs very significant.
From my experiment notebook its something like this :
exp no. | augmentation | pretrained | purchase_method | score_pretraining_phase | score_purchase_phase | score_validation_phase | LB_Score |
---|---|---|---|---|---|---|---|
1 | NO | NO | NO | 0.773 | 0.773 | 0.760 | |
2 | NO | NO | RANDOM 3000 | 0.773 | 0.804 | 0.760 | |
3 | NO | NO | ALL 10000 | 0.773 | 0.841 | 0.835 | |
4 | NO | YES | NO | 0.857 | 0.857 | 0.850 | |
5 | NO | YES | RANDOM 3000 | 0.857 | 0.864 | 0.845 | 0.851 |
6 | NO | YES | ALL 10000 | 0.857 | 0.892 | 0.875 | |
7 | YES | YES | NO | 0.868 | 0.868 | 0.865 | |
8 | YES | YES | RANDOM 3000 | 0.868 | 0.886 | 0.869 | 0.880 |
9 | YES | YES | ALL 10000 | 0.868 | 0.902 | 0.893 |
the notebook :
My Multiple Experiments Results ( the random one got 0.88 on LB)
Almost 3 years agoHereβs my multiple experiment results score that I log into tables.
Iβm using this same parameter for each experiment :
model : efficienet-b1
input: raw image
epoch: 20
optim: Adam
tl;dr, use augmentation and pre-trained weight.
I hope you it help you guys, especially for those who just joined.
Size of Datasets
Almost 3 years agowhatβs your comment about this @shivam ?
I think itβs 5000 training images, 3000 to purchase, 3000 to test right? just as in the overview.
Full list of available pretrained weights
Almost 3 years agowow, which pytorch version that got vit? nightly version?
Submit failed with no error log
Almost 3 years agoyes, it should be like that but mine didnβt show up.
What is this validation submission phase error log means?
Almost 3 years ago==================================
Deleting unsupported pre-trained model: ./.cache/pip/wheels/76/ee/9c/36bfe3e079df99acf5ae57f4e3464ff2771b34447d6d2f2148/gym-0.21.0-py3-none-any.whl
Deleting unsupported pre-trained model: ./.cache/pip/http/1/3/0/c/a/130ca645ced2b235e6f69505044bb4923f610dbb4bc6c8e1d76a50bb
Deleting unsupported pre-trained model: ./.cache/pip/http/8/f/8/e/b/8f8eb31d64d7424ab679aad519c22a7bf4f40ab17d1c4bad52b49a9c
Deleting unsupported pre-trained model: ./.cache/pip/http/a/d/c/0/3/adc03ed04ad13ffdeee3c838911d25a9f3659c9e3590f34fa6bf3a7e
Deleting unsupported pre-trained model: ./.git/objects/pack/pack-c148ae0f71d82068775278a3044e1a3c25b5f4a3.pack
Time left: 10800
timeout: the monitored command dumped core
/home/aicrowd/run.sh: line 38: 61 Segmentation fault timeout -s 9 $AICROWD_TIMEOUT_INFO python aicrowd_client/launcher.py
Submit failed with no error log
Almost 3 years agoI got error in the βValidate Submissionβ phase but with no log too
π Discussion on Starter Kit
Almost 3 years agohi @vrv ,
I tried using this submission method instead : AIcrowd
the push works, checked it on gitlab, but somehow itβs not on the submission. the tag `submission-`` prefix is right too. Any idea why?
Can we access the labels.csv files from the training data folder?
Almost 3 years agoIs it possible to access and read the csv files of training data labels?
or we only have the access to training_dataset class?
whatβs the filepath to access it? is it just ./data/training/images/labels.csv just like in the notebook example?
Allowance of Pre-trained Model
Almost 3 years agoIn my past experience doing challenges in Aicrowd, the committee has been really fair. I spotted some accounts that I suspected of cheating that before I reported it (with evidence of course), they have already taken care of it.
Seismic Facies Identification Challenge
Which average method is used in the calculation of f1-score?
About 4 years agoFrom the discord, @mohanty said itβs macro. But they might change it to weighted maybe in the 2nd round.
cmiiw
Notebooks
-
[Task 3 - Score 0.836] 3 Common Models Trained Separately Just 3 Common Model Trained SeparatelyleocdΒ· Over 2 years ago
-
Simple Way to Detect Noisy Label with opencv using opencv to enhance your strategy on training and buyingleocdΒ· Over 2 years ago
-
Ways to Select Which Data to Purchase - Episode 1 Active Learning MethodsleocdΒ· Almost 3 years ago
-
[LB 0.880] My Experiment Results + Baseline too I guess π¬ Experiment ResultsleocdΒ· Almost 3 years ago
-
Exploration, Body Paint Color Dominance, Image Transforms Data Exploration, Extracting Body Paint by Color Dominance from image, and Image TransformationleocdΒ· Almost 3 years ago
-
[Explainer] Introduction and General Approach Final Pack! Introduction to this challenge, general approach, my approach, and what I learn from the othersleocdΒ· Almost 4 years ago
-
[Explainer] Need extra features? Different input approach? Try Seismic Attributes! Basically itβs a math βInstagram-Snapchat-likeβ filter for seismic data. There are a lot of Seismic Attributes available.leocdΒ· About 4 years ago
I need to say this
Over 2 years agoWow. What a clickbait-y title. But that got your attention
I havenβt properly said it before, but Thank you Zew and Aicrowd for organizing this competition. Thank you, fellow participants. I learn lots of new stuff from this, especially from the top LB solutions & other participantsβ notebooks. I think I already got in my mind the best practice when facing this kind of problem in my work in the near future (sooner or later I think Iβll be facing this too, and labeling will be more expensive because of engineer/scientist level labeler needed for the data).
Hope you guys are always in good health.
Cheers