The workshop was the first since we started the cooperation with the budapest.py meetup group, and it was the 15th we’ve done since January, when we started to study together.
First Dóri talked about Anaconda and Jupyter notebooks, then showed us some basic pandas code for EDA (Exploratory Data Analysis) on a dataset about Pokemons. She walked through us how to get insights from our data, how to slice it, make new entries and lots more. After some basic visualization and aggregation we found out which Pokémon is the strongest based on our simple analysis. The second part was Berci’s and his Pandas tutorial notebook, which holds over a hundred different pandas functions and examples. He gave a short tour of the notebook. We aim to create some tutorial notebooks to help us focus on understanding the current dataset and spend less time looking up functions. These notebooks are going to help newcomers catch up with returning participants.
Solving this task at first I started with the standard Decision Tree, without any tuning. Then I get into GridSearchCV and RandomizedSearchCV for the best parameters. But after tweaking the model with these validations, I still couldn’t get higher than 79%. RandomForest didn’t help either.
That’s when I found XGBoost, a powerful model, getting more and more attention in machine learning. With it, I could go over 80%.