What’s in a name?

Johannes Stiehler
Cover Image for What’s in a name?

Some may claim that my co-founder Florian and I simply enjoy rebranding companies on a regular basis. To a certain extent, that’s true. “There’s magic in every beginning” and a new name and look capture that spirit of optimism. When we built the “ayfie” brand in 2016 - as a successor to the slightly ambiguous “VirtualWorks” - we also tried to do just that: Move away from enterprise search to a product for high-volume eDiscovery projects; move away from technology-centric to solution-centric messaging.

For our own brand, it is somewhat similar. Until recently, we were called “searchgears” in order to show that we provide all those little gear pieces needed to make a search solution truly great, in a manufactured quality and bespoke way: from front-end to data preparation, from “query suggest” to marketing. This focus appealed primarily to customers who had very clear ideas about what they wanted to do with their content and just needed technological support to turn that vision into reality.

However, in our discussions with current and future business partners, it is apparent again and again that this type of customer tends to be a minority. The majority is aware of having a problem with textual data, but does not yet have a clear idea of how to solve it. Search engines surprisingly often end up being part of such a solution, but never the only one and often not even the essential one.

In reality, this has long been our self-image, and now brand optics are following suit: We do whatever it takes to bring new movement (new motion) into big data piles. To find the needle in the haystack, you first have to get the haystack into shape. This is the goal of our products and services which we will be launching here in the next few months.

To pass the time, we will regularly disclose here what we have learned in 20 years in the “search business”.

Johannes Stiehler
CO-Founder NEOMO GmbH
Johannes has spent his entire professional career working on software solutions that process, enrich and surface textual information.

There's more where this came from!

Subscribe to our newsletter

If you want to disconnect from the Twitter madness and LinkedIn bubble but still want our content, we are honoured and we got you covered: Our Newsletter will keep you posted on all that is noteworthy.

Please use the form below to subscribe.

NEOMO is committed to protecting and respecting your privacy and will only use your personal data to manage your account and provide the information you request. In order to provide you with the requested content, we need to store and process your personal data. If you consent to us storing your personal data for this purpose, please check the box below.

Follow us for insights, updates and random rants!

Whenever new content is available or something noteworthy is happening in the industry, we've got you covered.

Follow us on LinkedIn and Twitter to get the news and on YouTube for moving pictures.

Sharing is caring

If you like what we have to contribute, please help us get the word out by activating your own network.

More blog posts


ChatGPT "knows" nothing

Language models are notoriously struggling to recall facts reliably. Unfortunately, they also almost never answer "I don't know". The burden of distinguishing between hallucination and truth is therefore entirely on the user. This effectively means that this user must verify the information from the language model - by simultaneously obtaining the fact they are looking for from another, reliable source. LLMs are therefore more than useless as knowledge repositories.


Rundify - read, understand, verify

Digital technology has overloaded people with information, but technology can also help them to turn this flood into a source of knowledge. Large language models can - if used correctly - be a building block for this. Our "rundify" tool shows what something like this could look like.


ChatGPT and the oil spill

Like with deep learning before, data remains important in the context of large language models. But this time around, since somebody else trained the foundation model, it is impossible to tell what data is really in there. Since lack of data causes hallucinations etc. this ignorance has pretty severe consequences.