Fashion trends can be difficult to predict — one minute something’s hot, the next it’s not at all. However, a team of researchers from IRI in Barcelona and the University of Toronto have built a mathematical model that will help us understand how something becomes fashionable.
The algorithm was presented as part of a paper entitled “Neuroaesthetics in Fashion: Modeling the Perception of Fashionability” at a conference held in Boston last month, according to a report from Science Daily. Edgar Simo-Serra and Francesc Moreno-Noguer from IRI teamed with Sanja Fidler and Raquel Urtasun from the University of Toronto to make the project a reality.
The model was built using a dataset of fashion-themed posts found on the Internet. Images of fashionable clothing and the captions they were given were cross-referenced against the social media response to each entry. Individual garments were judged as fashionable or not fashionable based on the amount of likes they received on social media.
Basically, the method boils down to posting a picture of an outfit on Facebook and seeing whether people love it or hate it — which is, in of itself, a decent summation of how fashion works. However, the project can claim some authority on what to wear thanks to the scale of the sample being used, with data from over 144,000 posts being taken into consideration.
The intention is that the system will be able to advise users on fashion choices thanks to the data it has accumulated. What remains to be seen is whether anyone will subject themselves to having a computer give them tips on how to dress in the morning. If the idea of fashion can really be boiled down to a statistical algorithm, wouldn’t we all end up wearing the exact same outfit?
Editors' Recommendations
- These are the PC apps you’re still using — but probably shouldn’t be
- ChatGPT AI chatbot can now be used without an account
- If you’re still gaming on 16GB of RAM, you’re missing out
- OpenAI’s new tool can spot fake AI images, but there’s a catch
- Meta is reportedly working on a GPT-4 rival, and it could have dire consequences