Seite wählen

NETWAYS Blog

OSMC 2023 | Experiments with OpenSearch and AI

Last year’s Open Source Monitoring Conference (OSMC) was a great experience. It was a pleasure to meet attendees from around the world and participate in interesting talks about the current and future state of the monitoring field.

Personally, this was my first time attending OSMC, and I was impressed by the organization, the diverse range of talks covering various aspects of monitoring, and the number of attendees that made this year’s event so special.

If you were unable to attend the congress, we are being covering some of the talks presented by the numerous specialists.
This blog post is dedicated to this year’s Gold Sponsor Eliatra and their wonderful speakers Leanne Lacey-Byrne and Jochen Kressin.

Could we enhance accessibility to technology by utilising large language models?

This question may arise when considering the implementation of artificial intelligence in a search engine such as OpenSearch, which handles large data structures and a complex operational middleware.

This idea can also be seen as the starting point for Eliatra’s experiments and their findings, which is the focus of this talk.

 

Working with OpenSearch Queries

OpenSearch deals with large amounts of data, so it is important to retrieve data efficiently and reproducibly.
To meet this need, OpenSearch provides a DSL which enables users to create advanced filters to define how data is retrieved.

In the global scheme of things, such queries can become very long and therefore increase the complexity of working with them.

What if there would be a way of generating such queries by just providing the data scheme to a LLM (large language model) and populate it with a precise description of what data to query? This would greatly reduce the amount of human workload and would definitely be less time-consuming.

 

Can ChatGPT be the Solution?

As a proof-of-concept, Leanne decided to test ChatGPT’s effectiveness in real-world scenarios, using ChatGPT’s LLM and Elasticsearch instead of OpenSearch because more information was available on the former during ChatGPT’s training.

The data used for the tests were the Kibana sample data sets.

Leanne’s approach was to give the LLM a general data mapping, similar to the one returned by the API provided by Elasticsearch, and then ask it a humanised question about which data it should return. Keeping that in mind, this proof of concept will be considered a success if the answers returned consist of valid search queries with a low failure rate.

 

Performance Analysis

Elasticsearch Queries generated by ChatGPT (Result Overview)

Source: slideshare.net (slide 14)

As we can see, the generated queries achieved only 33% overall correctness. And this level was only possible by feeding the LLM with a number of sample mappings and the queries that were manually generated for them.

Now, this accuracy could be further improved by providing more information about the mapping structures, and by submitting a large number of sample mappings and queries to the ChatGPT instance.
This would however result in much more effort in terms of compiling and providing the sample datasets, and would still have a high chance of failure for any submitted prompts that deviate from the trained sample data.

 

Vector Search: An Abstract Approach

Is there a better solution to this problem? Jochen presents another approach that falls under the category of semantic search.
Large language models can handle various inputs, and the type of input used can significantly impact the results produced by such a model.
With this in mind, we can transform our input information into vectors using transformers.
The transformers are trained LLM models that process specific types of input, for example video, audio, text, and so on.
They generate n-dimensional vectors that can be stored in a vector database.
Illustration about the usage of vector transformers

Source: slideshare.net (slide 20)

When searching a vector-based database, one frequently used algorithm for generating result sets is the ‚K-NN index‘
(k-nearest-neighbour index). This algorithm compares stored vectors for similarity and provides an approximation of their relevance to other vectors.
For instance, pictures of cats can be transformed into a vector database. The transformer translates the input into a numeric, vectorized format.
The vector database compares the transformed input to the stored vectors using the K-NN algorithm and returns the most fitting vectors for the input.

 

Are Vectors the Jack of all Trades?

There are some drawbacks to the aforementioned approach. Firstly, the quality of the output heavily depends on the suitability between the transformer and the inputs provided.
Additionally, this method requires significantly more processing power to perform these tasks, which in a dense and highly populated environment could be the bottleneck of such an approach.
It is also difficult to optimize and refine existing models when they only output abstract vectors and are represented as black boxes.
What if we could combine the benefits of both approaches, using lexical and vectorized search?

 

Retrieval Augmented Generation (RAG)

Retrieval Augmented Generation (RAG) was first mentioned in a 2020 paper by Meta. The paper explains how LLMs can be combined with external data sources to improve search results.
This overcomes the problem of stagnating/freezing models, in contrast to normal LLM approaches. Typically, models get pre-trained with a specific set of data.
However, the information provided by this training data can quickly become obsolete and there may be a need to use a model that also incorporates current developments, the latest technology and currently available information.
Augmented generation involves executing a prompt against an information database, which can be of any type (such as the vector database used in the examples above).
The result set is combined with contextual information, for example the latest data available on the Internet or some other external source, like a flight plan database.
This combined set could then be used as a prompt for another large language model, which would produce the final result against the initial prompt.
In conclusion, multiple LLMs could be joined together while using their own strengths and giving them access to current data sources and that could in turn generate more accurate and up to date answers for user prompts.
Overview of the RAG (Retrieval Augmented Generation)

Source: slideshare.net (slide 36)

Noé Costa
Noé Costa
Developer

Noé ist als Schweizer nach Deutschland ausgewandert und unterstützt das Icinga Team seit Oktober 2023 als Developer im Bereich der Webentwicklung. Er wirkt an der Weiterentwicklung der Webmodule von Icinga mit und ist sehr interessiert am Bereich des Monitorings und dessen Zukunft. Neben der Arbeit kocht er gerne, verbringt Zeit mit seiner Partnerin, erweitert sein Wissen in diversen Gebieten und spielt ab und an auch Computerspiele mit Bekanntschaften aus aller Welt.

OSMC 2024 is Calling for Sponsors

What about positioning your brand in a focused environment of international IT monitoring professionals? Discover why OSMC is just the perfect spot for it.

 

Meet your Target Audience

Sponsoring the Open Source Monitoring Conference is a fantastic opportunity to promote your brand.
Raise corporate awareness, meet potential business partners, and grow your business with lead generation. Network with the constantly growing Open Source community and establish ties with promising IT professionals for talent recruitment. Connect to a diverse and international audience including renowned IT specialists, Systems Administrators, Systems Engineers, Linux Engineers and SREs.

 

Your Sponsorship Opportunities

Our sponsorship packages are available in a variety of budgets and engagement preferences: Platinum, Gold, Silver, and Bronze.
From an individual booth, speaking opportunities and lead scanning to social media and logo promotion in different ways, everything is possible for you.
We additionally offer some Add-Ons which can be booked separately. Use this unique chance to get even more out of it. Sponsor the Dinner & Drinks event, the Networking Lounge or the Welcome Reception.

Download the sponsor prospectus for full details and pricing.

We look forward to hearing from you!

 

Early Bird Alert

Our Early Bird ticket sale is already running. Make sure to save your seat at the best price until May 31.
Our discounted tickets are selling fast, grab yours now before they’re gone!

 

Save the Date

OSMC 2024 is taking place from November 19 – 21, 2024 in Nuremberg. Mark your calendars and be part of the 18th edition of the event!

Katja Kotschenreuther
Katja Kotschenreuther
Manager Marketing

Katja ist seit Oktober 2020 Teil des Marketing Teams. Als Manager Marketing kümmert sie sich hauptsächlich um das Marketing für die Konferenzen stackconf und OSMC sowie unsere Trainings. Zudem unterstützt sie das Icinga Team mit verschiedenen Social Media Kampagnen und der Bewerbung der Icinga Camps. Sie ist SEO-Verantwortliche für all unsere Websites und sehr viel in unserem Blog unterwegs. In ihrer Freizeit reist sie gerne, bastelt, backt und engagiert sich bei Foodsharing. Im Sommer kümmert sie sich außerdem um ihren viel zu großen Gemüseanbau.

OSMC 2023 | Will ChatGPT Take Over My Job?

One of the talks at OSMC 2023 was „Will ChatGPT take over my job?“ by Philipp Krenn. It explored the changing role of artificial intelligence in our lives, raising important questions about its use in software development.

 

The Rise of AI in Software Development:

In a time of technological advancement, discussing the use of artificial intelligence, represented by tools like ChatGPT, has become unavoidable. The younger generation, eager to embrace AI, faces a pivotal question: will these advanced language models replace traditional coding methods or serve as powerful enhancements?

 

Insights from Philipp’s Presentation:

Philipp shared some insights from a SauceLabs survey, revealing that over a quarter of developers admitted to neglecting proper testing. This statistic reveals possible issues with the trustworthiness of AI-generated code, emphasizing the need for careful testing.

 

Navigating the Pitfalls:

Relying solely on ChatGPT has risks. The changing data and the model’s tendency to alter responses make it unpredictable. Accepting AI solutions without thorough validation endangers code integrity.

 

Practical Application in Debugging:

Drawing on his experience at Elastic, Philipp demonstrated a practical use case where an AI Assistant aids in explaining error messages. While traditional debugging methods remain crucial, the integration of AI simplifies and enhances this process. The demonstration showcased the assistant’s capability to execute functions, providing refined and efficient solutions to coding queries. You can also watch the recording of Philipp’s talk and check out the detailed demonstration.

 

The Path Forward:

In conclusion, ChatGPT and similar AI models undeniably transform the software development landscape. It’s crucial to have a balanced view, avoiding blind trust in their capabilities. As we adapt to this changing phase, integrating AI into our workflows should be done carefully. While ChatGPT won’t completely replace our roles, it significantly influences the future of software development. Embracing this evolution involves a mix of trust, validation, and a strategic approach to the evolving technological landscape. I think at the moment it may be a bit too early to trust ChatGPT 100%, but we should definitely learn how to use it.

Make sure to watch the video of his recorded talk and have a look at Philipp’s slides.

I’m already looking forward to OSMC 2024 and I hope to see you there!

Jonada Hoxha
Jonada Hoxha
Developer

Nach erfolgreichem Abschluss ihrer Ausbildung als Fachinformatikerin für Anwendungsentwicklung verstärkt Jonada das Development unseres Icinga Teams. In ihrer Freizeit ist sie entweder in der Turnhalle aktiv oder vertieft sich in Computerspielen am PC. Aktuell ist ihr Ziel in Apex Legends, den Rank "Apex Predator" zu erreichen.

stackconf 2024 | Announcing our First Speakers

The Wait is Over!

We’re super excited to finally announce our first confirmed speakers for stackconf 2024. It’s a diverse group of international experts who specialize in cloud-native infrastructure solutions. Have a look and see which topics interest you the most. There are more to come in the next few weeks. Stay tuned!

 

Spoiler Alert!

Here’s a small insight into our great line-up:

ALEX PSHE

Step-by-step algorithm for building CI/CD as an automated quality control system

Alex’s presentation delves into CI/CD from a tester’s perspective, highlighting the importance of leveraging automated metrics for decision-making and adopting a fail-first approach in pipeline design. Throughout the talk, she explores different pipeline types and triggers for efficient testing, emphasizing key criteria for effectiveness. The role of Quality Gates and automated control systems in enforcing standards is also discussed, underscoring Alex’s focus on ensuring reliability and quality in software delivery.

 

ARNOLD BECHTOLDT

Building large-scale Internal Developer Platforms: Core Principles

Arnold’s talk navigates the complexities of cloud computing, transcending its enigmatic allure. He highlights RTL Data’s internal developer platform, serving 30+ teams with diverse technologies. Arnold underscores RTL’s decentralized infrastructure, which has rapidly produced over 400 GCP projects and 100 GKE clusters in 2-3 years, showcasing its commitment to scalable and efficient software delivery.

 

ANAÏS URLICHS

Looking into the Closet from Code to Cloud with Bills of Material

Anaïs talks about Bills of Material (BOM) and how they are changing. In her presentation, she shows how to deal with differences in tool and BOM outputs. Using easy-to-understand demos, she explains how to make BOMs with Trivy, Syft, and Microsoft SBOM. Anais focuses on checking security and quality with Lockheed Martin’s sbom-comparator. She also shows how to use BOMs in security scans and reduce vulnerability scan noise with VEX documents.

 

VISHWA KRISHNAKUMAR

Scaling Up, Not Out: Managing Enterprise Demands in a Growing SaaS Startup

Every founder reaches a crucial point when their product gains traction. They attract big clients, like software leaders, banks, and retail giants, who pay generously for niche solutions. While exciting, catering to these clients‘ demands can veer the product away from its original vision. Vishwa’s talk advises founders on balancing enterprise customer needs while staying true to their company’s vision.

 

Save your Ticket!

Curious about the event program and excited to join stackconf 2024 on June 18 & 19 in Berlin? Act now and reserve your spot! Make sure you don’t miss any important stackconf updates and sign up for our mailing list. Stay connected for upcoming announcements!

 

Katja Kotschenreuther
Katja Kotschenreuther
Manager Marketing

Katja ist seit Oktober 2020 Teil des Marketing Teams. Als Manager Marketing kümmert sie sich hauptsächlich um das Marketing für die Konferenzen stackconf und OSMC sowie unsere Trainings. Zudem unterstützt sie das Icinga Team mit verschiedenen Social Media Kampagnen und der Bewerbung der Icinga Camps. Sie ist SEO-Verantwortliche für all unsere Websites und sehr viel in unserem Blog unterwegs. In ihrer Freizeit reist sie gerne, bastelt, backt und engagiert sich bei Foodsharing. Im Sommer kümmert sie sich außerdem um ihren viel zu großen Gemüseanbau.

OSCamp 2024 | Das Programm ist online!

Tauche ein in die Welt unserer Referenten

Wir freuen uns, mitteilen zu können, dass das Programm des Open Source Camps für Kubernetes feststeht! Wir haben eine vielfältige Auswahl an Vorträgen zusammengestellt, die Dir einen umfassenden Einblick in die neuesten Kubernetes Lösungen bieten.

Unsere Speaker sind alle Experten auf ihrem Gebiet und bringen jede Menge praktische Erfahrungen mit sich. Sie werden über die neuesten Entwicklungen, Trends und bewährte Praktiken im Kubernetes-Bereich sprechen und Dir wertvolle Einblicke geben.

 

Achtung Spoiler!

Unter anderem erwarten Dich diese Speaker mit ihren Vorträgen.

 

Alex Pshe

Step-by-step algorithm for building CI/CD as an automated quality control system (EN)

Wir alle wissen, wie wichtig automatische Bereitstellungen von Anwendungen sind. Doch wissen wir auch, wie man ein automatisiertes Qualitätskontrollsystem aufbaut? Alex hat einen Schritt-für-Schritt-Algorithmus für die Einrichtung eines solchen Systems für Webanwendungen vorbereitet, einschließlich bewährter Praktiken von Linting bis zur Leistungstestung.

 

 

Andreia Otto & Ravi Mogulla

SRE Challenges in Monolith to Microservices Shift at adidas E-commerce (EN)

Andreia und Ottos Vortrag gewährt Einblicke in die Umstellung von adidas E-Commerce auf Microservices und die damit verbundenen Herausforderungen für SREs. Dabei werden Themen wie Skalierbarkeit, Bereitstellung, Überwachung und Vorfälle ausführlich behandelt. Erlerne praktische Lektione und entdecke inspirierende Erfolgsgeschichten, die Dir helfen, Dich in der sich ständig wandelnden Landschaft des Site-Reliability-Engineerings (SRE) im Zeitalter der Microservices zurechtzufinden.

 

 

Magnus Kulke

Confidential Containers – Sensible Daten und Privatsphäre in Cloud Native Umgebungen (DE)

Magnus Vortrag beleuchtet die Verwendung von Confidential Containers für sensible Daten in Cloud-Native-Umgebungen, mit Fokus auf grundlegenden Konzepten von Confidential Computing und ihrer Umsetzung in Linux und Hardware. Er stellt zudem das „Confidential Containers“ Projekt der „Cloud-Native Computing Foundation“ vor, gefolgt von einer praktischen Demo.

 

Isabelle Rotter

What gets measured gets communicated (DE)

Projekte sollten Spaß machen – beim Codieren und bei der Kommunikation. Isabelles Vortrag zeigt, wie Kommunikation in technischen Teams verbessert werden kann, basierend auf Erkenntnissen aus der Psychologie. Sie bietet eine Schritt-für-Schritt-Anleitung, um die Kommunikation effektiver und unterhaltsamer zu gestalten.

 

 

 

Alex König

Running WebAssembly on Kubernetes (EN)

WebAssembly hat das Potenzial, die Cloud-Native Anwendungsentwicklung zu revolutionieren. Im Vergleich zu Containern bietet es Vorteile wie bessere Portabilität, stärkere Isolation und schnellere Startzeiten. Alex zeigt Dir in seinem Vortrag, wie Du WebAssembly in Deinen Kubernetes-Cluster integrieren und Workloads effektiv ausführen, skalieren und orchestrieren kannst.

 

Jetzt Ticket sichern!

Wir laden Dich herzlich ein, am 18. April in Nürnberg mit dabei zu sein. Vernetze Dich mit anderen Kubernetes-Begeisterten, tausche Dich aus über die neuesten Entwicklungen und erweitere Dein Wissen. Hol Dir jetzt Dein Ticket! Wir freuen uns auf Dich!

 

Katja Kotschenreuther
Katja Kotschenreuther
Manager Marketing

Katja ist seit Oktober 2020 Teil des Marketing Teams. Als Manager Marketing kümmert sie sich hauptsächlich um das Marketing für die Konferenzen stackconf und OSMC sowie unsere Trainings. Zudem unterstützt sie das Icinga Team mit verschiedenen Social Media Kampagnen und der Bewerbung der Icinga Camps. Sie ist SEO-Verantwortliche für all unsere Websites und sehr viel in unserem Blog unterwegs. In ihrer Freizeit reist sie gerne, bastelt, backt und engagiert sich bei Foodsharing. Im Sommer kümmert sie sich außerdem um ihren viel zu großen Gemüseanbau.