13 posts with this tag
Over the past two years, since the initial release of ChatGPT in November 2022, we've been in a hype cycle around large language models, or LLMs. Due to the very human "feeling" of LLM outputs, a lot of people have been convinced that these models …
With a full coronavirus shutdown in Germany, Christmas this year in Berlin will be very quiet. I have been keeping myself entertained for the last couple of weeks by watching old seasons of The Great British Bakeoff, so between that and the mountain …
In this post, we continue our discussion about how to use AWS Sagemaker's BlazingText to train a word2vec model. In the last post we learned how to set up, train and evaluate a single model. However, we essentially selected our hyperparameters at …
AWS Sagemaker has a number of inbuilt algorithms, which are not only easier to use with the Sagemaker set up but are also optimised to work with AWS architecture. At my previous job, we used word embeddings extensively to help solve NLP problems. We …
One of the most basic (and most important) tasks when doing text mining is cleaning up your text. While this might seem a bit dull compared to sexy stuff like sentiment analysis and topic modelling, I hope to show you in this post that not only is …