AWS for Industries

Boost automotive productivity with process automation facilitated by Generative AI

Enterprises today are continuing to explore new solutions to increase efficiency, particularly around process automation. That efficiency is pertinent to multiple industries, but it is especially notable in the automotive sector. Automotive research and development (R&D) teams need to address costly, time-consuming, and error-prone processes, for example, in generating request for proposal (RFP) documents.

The development cycle for new vehicles is complex and can last for up to 5 years. During that time, thousands of components—such as infotainment modules—are designed, built, and tested. Part of the work of the development cycle is conducted by the original equipment manufacturer (OEM)—that is, the vehicle-producing company—but large portions of the effort are solved in collaboration with suppliers.

Automotive companies often use RFPs when awarding various types of work to vendors. The use of RFPs empowers automotive companies to use the expertise and capabilities of external vendors while also maintaining control over the selection process and making sure that chosen vendors align with relevant requirements and objectives.

Based on a given RFP, potential suppliers generate offers detailing their approaches through a two-step process. First, the suppliers’ technical approaches are evaluated by the department that issued the RFP. If a bid is approved by the issuing department, the OEM’s procurement team reviews the price. Depending on the number of applicants, the process can last anywhere from several weeks to months before the final suppliers are selected.

In theory, an initial RFP should contain all required information and context for the request, including fully defined technical specifications. But in practice, an RFP may contain only a short brief with vague descriptions. RFPs that are not issued under a strict process may miss critical information derived from previous projects and overall historical knowledge.

To address those challenges, Accenture, Amazon Web Services (AWS), and Intel have collaborated on an RFP generation solution using generative artificial intelligence (AI). In considerably less time, the RFP generator automatically gathers critical company information and pre-populates content in the RFP. Users then receive automated feedback thanks to a verbal exchange with the knowledge base, empowering them to focus on critical tasks elsewhere. Improved RFP quality streamlines supplier assessment by eliminating the need for back-and-forth communication with suppliers, making the entire process more efficient.

In this blog, we dive into specific features and benefits of the RFP generator, which is powered by Intel-based instances for Amazon Elastic Compute Cloud (Amazon EC2)—secure and resizable compute capacity for virtually any workload—and Intel® Tiber™ AI Studio.

Solution overview

The solution had to match a few essential parameters. First, it needed to augment the capabilities of automotive procurement employees, meaning that it needed to complement existing RFP-generation processes as opposed to fully replacing human-fulfilled tasks. Second, the tool had to use generative AI to deliver significant performance improvement at a price point that would make deployment at scale financially viable.

At the solution’s core, we put Intel® Tiber™ AI Studio, a cloud-agnostic machine learning platform. Intel Tiber AI Studio has a container-based architecture that helps facilitate seamless deployment across multiple environments, supporting hybrid cloud strategies and ease of migrating workloads. That adaptability extends to the creation of heterogeneous compute pipelines, where different stages can run on optimized architectures, such as preprocessing on CPUs and training on GPUs or HPUs, regardless of the provider.

We deployed Intel Tiber AI Studio using different AWS services. For orchestration, we used Amazon Elastic Kubernetes Service (Amazon EKS), a managed Kubernetes service to run Kubernetes in the AWS cloud. For DNS management, we used Amazon Route 53, a reliable and cost-effective way to route end users to internet applications. We integrated worker nodes from Intel® Tiber™ Developer Cloud, such as 4th Gen Intel® Xeon® Scalable Processor and Intel® Gaudi® 2 AI accelerators from Habana Labs, and served our self-hosted large language models (LLMs) as API endpoints through Intel Tiber AI Studio. That arrangement empowered us to successfully integrate the endpoints into the RFP generation and ticket-creating process.

Moreover, we delivered this solution using a self-hosted LLM (Mistral 7B) that is supported by Accenture and Intel’s AI Playground.

The details

Figure 1. RFP context diagram

Figure 1. RFP context diagram

Figure 1 illustrates the solution’s workflow for generating RFPs. The process begins with user input through a Streamlit-based frontend interface. That data is then processed by a Python backend for prompt engineering, making sure that the input is prepared for further processing. The backend connects with an endpoint to Intel® Tiber™, where a Kubernetes web service is deployed. Key elements of the solution are as follows:

Cache check: Helps verify if the same query was sent previously.

Retrieval query: Augments the prompt with relevant data.

PDF/Word export: Generates an RFP step by step, reviewing each document subsection sequentially.

The solution enforces a specific structure by chapter and subsection to generate one section at a time. That means the solution can provide individual instructions for each section and establish a standard for writing future RFPs.

Again, the challenge was not how to automate the process completely but how to create a tool that would more seamlessly complement an employee’s own process for creating context-specific RFPs—in as little time as possible. To do that, we used a self-hosted Mistral 7B model that met accuracy needs with a sufficiently large context window, a robust MLOps tool, and vector storage to house past RFPs.

Our model selection criteria were set according to a range of factors around data security, cost, and ability to handle a sufficiently large context window. Of utmost necessity was the incorporation of an open source model so that we could control deployment scenarios and provide greater data security by avoiding any dependency on a third-party provider. We looked at a variety of model sizes, and with this, the cost of deployment and selection of hardware to support the model endpoint were key considerations. We spent time comparing models as large as Llama 2 70B and models as small as 7–8 billion parameters. When we looked at the effectiveness of responses generated through our RAG pipeline, we found that we could get a high quality of response from a smaller model capable of handling a sufficiently large context window. That requirement was met by Mistral 7b, which we ultimately selected. Thus, in the end, a smaller model met our needs while being more cost-effective.

Key decisions

In sum, the key decisions that we made were as follows:

LLM selection: We selected Mistral 7B over other models in view of factors like greater security, lower cost, and ability to handle sufficiently large context. We spent time comparing other models, such as Llama 2 70B, but ultimately a smaller model met our needs while being more cost-effective.

Performance optimization: We used Intel Extension for PyTorch to help increase inference speed on Intel hardware.

Cross-cloud model for development and deployment: We used Intel Tiber AI Studio to stitch together workflows that might use components in different clouds (for example, fine-tuning an Intel Gaudi 2 AI accelerator in the Intel Tiber Developer Cloud) while still having our model-serving inference requests done on AWS.

Customer benefits

Model ownership and security: Keeping data in a “walled garden” helps increase overall security.

Time and cost savings: Potential time and cost reduction in RFP generation contributes to a more streamlined research and decision-making process.

Scalability and consistency: AI-powered solutions can better handle increasing volumes of RFPs and error tickets without compromising on quality.

Improved accuracy and compliance: The ability to integrate historical data and industry regulations reduces human error and helps customers meet automotive industry standards.

Faster time to market: Automating administrative tasks can accelerate the overall R&D process, potentially leading to faster product development and time to market of new automotive technologies.

Conclusion

Our pilot customer benefited from an industry-leading solution that streamlined its work process without introducing net-new AI costs that might have offset gains in productivity. By adopting the solution, the automotive manufacturer can stay ahead of its competition and meet growing industry demand.

Accenture’s generative AI solutions, powered by AWS and Intel technologies, offer improved efficiency and cost savings in automotive environments. The three companies have a long history of collaborating to drive business transformation through advanced technology solutions, which have spanned AI, cloud transformation, and edge computing. Our custom engagement makes use of Accenture’s deep industry expertise, Intel’s hardware and software innovation, and the AWS cloud and AWS services to provide a seamless customer solution. Reach out to us if you’d like to learn more.

Accenture’s AWS Partner Profile | Contact Accenture | Intel’s AWS Partner Profile | Contact Intel

Nick Huntley

Nick Huntley

Nick has been with Intel for 15 years, currently leading the EMEA alliance between Intel and Accenture's AI initiatives from his UK base. In this role, Nick focuses on collaboratively developing "get-to-market" and "go-to-market" strategies with Accenture, driving impactful value for mutual clients across the EMEA region. His work also involves fostering innovation and building strategic ecosystem partnerships that align technology solutions with business outcomes.

Dr. Christian Kohlschein

Dr. Christian Kohlschein

Christian works at Accenture’s Industry X division, where he and his team focus on the development of industrial end-to-end data analytics solutions, with a focus on the manufacturing industry, especially automotive. He has extensive project experience in the successful implementation and management of AI-based projects. Before joining Accenture, he was co-founder and technical director of an AI-powered data analytics company in the manufacturing sector. Prior to that, he was managing senior engineer of a large mechanical engineering department at RWTH Aachen University and software developer in the business intelligence area of the aviation industry. During his studies and doctorate, he specialized in AI and the associated ecosystem of technologies, people and processes.

Colin Best

Colin Best

Colin is a Data Scientist in Accenture's Data & AI practice, focused on using machine learning and generative AI to solve business challenges. He is the Enterprise AI Lead for the Accenture and Intel Partnership where he focuses on emerging innovation and the optimization of AI workloads.

Soonam Kurian

Soonam Kurian

Soonam Kurian is a Principal Solutions Architect at Amazon Web Services specializing on partner solutions. She is focused on data analytics, artificial intelligence, and machine learning. In her current role, she works with Global System Integrators and Independent Software Vendors providing architectural guidance to launch strategic industry solutions on AWS. In her spare time, she enjoys reading and being outdoors.