Key Takeaways
|
|
Technical developments surrounding generative artificial intelligence (GenAI) are advancing at blazing speed and are beginning to change our world for the better. This makes it all the more important to recognize the potential of this technology and leverage it sensibly to increase your competitiveness. Automic Automation is an automation platform that optimally equips you for the AI revolution—combining proven enterprise capabilities with the mind-blowing potential of GenAI.
Automation platforms have always been an indispensable enabler of digital transformation and, as such, are ideally suited to channeling AI into orderly processes and harnessing it effectively. Automic Automation and GenAI prove to be an ideal combination on several levels: GenAI provides Automic users with helpful support in implementing new automation projects and it facilitates the smooth operation of automation processes already running in production. In addition, GenAI can also function as an integral part of automation workflows, enriching deterministic tasks with intelligent decision-making and enabling the development of entire workflows as AI agents.
How retrieval-augmented generation (RAG) improves GenAI output
Despite the rapid development of new large language models (LLMs), their inherent knowledge is limited at the time of their release. For example, an LLM published a week ago knows nothing about the new scientific study published yesterday or the latest version of a REST API released just the day before. Nor would it have visibility into the recently published knowledge base article or the latest update to the corporate travel policy. To harness the potential of GenAI, such information must be integrated into the context of a conversation with an LLM. The size of the context itself, the so-called context window, is also more or less limited depending on the model chosen. Although the technical limits are increasingly being expanded, the number of tokens continues to play at least a commercial role as a common metric for billing for corresponding services.
A common approach to solving this problem is called "retrieval-augmented generation" (RAG). This method is used to identify relevant information from specific knowledge sources (retrieval), which is then integrated into the context of a query to an LLM (augmentation), thus enabling the LLM to produce a more accurate answer (generation). Instead of integrating all knowledge base articles into the context, the relevant parts of selected knowledge base articles are first identified based on the query to the LLM, and only these are used. By including only the truly relevant information in the request, the number of tokens consumed can be drastically reduced. Further, the accuracy and relevance of the response can also be significantly improved, and the likelihood of hallucinations can be reduced.
Automic Automation fuels RAG
But how does such a system work in practice? Let's imagine the following scenario: At regular intervals, the contents of new documents must be transferred to a vector database, enabling users to employ an LLM to access the contents of their company documents. Before this, a few additional steps are usually necessary. For example, the text must first be extracted from PDF documents and then broken down into smaller pieces using special splitter algorithms, which can take overlaps into account to maintain semantic coherence. In the next step, vector representations of the individual text blocks are generated using a so-called embedding model. These blocks are then stored in a vector database. Upon closer inspection, this is a classic use case for automation. And indeed, such a "RAG pipeline" can be implemented quite easily with Automic Automation.
Automic’s low-code approach makes your life easier
First, you need to identify the data source and check the Automation Marketplace for the availability of a suitable integration with Automic Automation. Broadcom offers a wide range of integrations with third-party applications, which makes the creation of automated processes much easier, as they eliminate the need for specialized knowledge of third-party APIs.
With the help of a suitable integration, the desired documents can be downloaded. Further processing of these documents is most easily accomplished with Python. In effect, there’s no way for data scientists to get around Python. This reality is reflected in the large selection of useful libraries. Automic Automation offers a new job type for this, the Python job. With this, Python becomes a first-class citizen. Automic Automation developers can rely on Python and popular libraries, such as Apache Tika and Langchain, to implement those parts of the RAG pipeline that are responsible for extracting text from the PDFs and splitting it into overlapping text blocks.
Automic REST jobs also prove helpful, allowing REST API requests to be configured in a snap. In addition, the API’s responses can be evaluated for further processing. For example, models made available via the Ollama platform can be used to generate vector embeddings using the Ollama REST API. Creating a corresponding Automic REST job makes this step much easier. The Automic REST job can also be used to store the vector representation of the individual text blocks in any vector database that also provides a corresponding REST API, such as Chroma DB. Ultimately, the individual jobs are linked together in an Automic workflow, thus automating the RAG pipeline.
RAG in action
Implementing the RAG Pipeline in combination with LLMs opens up entirely new possibilities. With the new ASK_AI Automic script function, conversations with an LLM become an integral part of an Automic workflow. In this way, individual prompts can also be used for a similarity search in documents stored in the vector store. Relevant text blocks can be identified and added to the context of the conversation with the LLM. In this application scenario, a REST job can be used to initiate the similarity search. The result of this is ultimately sent to the configured LLM in an Automic automation script with the ASK_AI script function.
Automic combines cutting-edge technologies with proven capabilities
Automic Automation proves to be an ideal platform for implementing such scenarios. The wide range of integrations available on the Automation Marketplace allows for easy connection to third-party applications. The data from these applications can ultimately be made available in a vector store. Thanks to Automic Automation's low-code approach, individual building blocks can be developed as jobs and ultimately combined into an Automic workflow. Specialized job types are available, and thanks to Python jobs, more complex business logic can also be programmed and thus integrated into the automation workflow.
Ultimately, Automic Automation also impresses customers with its proven capabilities for scheduling regular executions and its reliability in daily operations. GenAI and common design patterns such as RAG can be easily implemented with Automic Automation. It is therefore clear that Automic Automation is the right choice for combining cutting-edge technologies with proven capabilities, so teams can meet the challenges of the future.
For more on Automic Automation’s recently released features, read the following blog: Democratizing Automation With the AI-Powered Filter Assistant and Seamless Python Scripting.

Michael Grath
Michael leads the Broadcom AOD Automation engineering organization, which has over 150 global team members. His responsibilities include Automic Automation, AutoSys, and Automation Analytics & Intelligence products and promoting innovation in product development through the use of generative AI.
Other resources you might be interested in
Powering RAG Pipelines With Automic Automation
See how Automic Automation optimally equips you for the AI revolution, combining proven enterprise capabilities with the potential of generative AI.
Unlock Real-Time AWS Observability With Streaming Ingestion in DX Operational Observability
With streaming ingestion capabilities, DX Operational Observability offers visibility into your AWS telemetry, enhancing insights and incident response.
Observability and IT Monitoring Governance: Establishing Order (Part 3 of 4)
Find out how DX Unified Infrastructure Management (DX UIM) supports monitoring governance, enabling teams to manage configurations and track alarm policies.
Observability and IT Monitoring Governance (Part 4 of 4)
This post shows how baselines, KPIs, and thresholds are essential for monitoring governance. See how IT can shift from reactive to proactive IT management.
What's Really Happening in Your Branch Office Network?
Fragmented monitoring tools create critical visibility gaps in branch networks. Find out why you need network observability to pinpoint the cause of issues.
Rally Office Hours: August 28, 2025
Learn about the general availability of the AI writing assistant in Rally, then follow the weekly Q&A session with Rally product experts.
AI as Talent: Navigating the New Landscape of Artificial Intelligence Management
Adopt a strategic talent management approach to AI, so you can ensure investments generate sustainable value, improve competitive edge, and mitigate risks.
Automic Automation Cloud Integration: Google Dataproc Agent Integration
Broadcom's Google Dataproc Automation Agent lets you easily execute Dataproc jobs, monitor and manage them with your existing enterprise workload automation, as well as other cloud-native activities.
Observability and Monitoring Governance (Part 1 of 4)
Find out how strong monitoring governance can help IT teams cut through the noise, see what truly matters, and act with precision.