PaulHoule found these results unsurprising due to the noisy nature of the recommender and the unpredictability of user behavior on Hacker News. He revealed that his best model for these tasks remains a bag-of-words model, suggesting its superior efficiency and effectiveness in this context.
Key takeaways:
- PaulHoule tried fine-tuning BERT models for a content-based recommender system but found they required more than 50x the effort to train than an embedding based model.
- The performance of the BERT models and the embedding based model was approximately the same.
- Fine-tuned BERT was unsuccessful at predicting if a headline would get upvoted on HN or what the comments/votes ratios would be.
- The best model for predicting upvotes and comments/votes ratios on HN, according to PaulHoule, is still a bag-of-words model.