Softserve Four Adoption Patterns For Your Enterprise
Softserve Four Adoption Patterns For Your Enterprise
innovation
PATTERNS FORwith
YOUR
Generative
ENTERPRISE AI
The Future
is Now
Pundits have called generative AI technology a once-in-a-lifetime SoftServe’s adoption patterns
moment — a transformative technology to rival the commercialization aim to provide you — business
of the Internet. The hype is deafening, even confusing.
and Use
Your Apps
Your Data
Your Data
Rapidly adopt third-party Seamlessly integrate Customize third-party Develop custom Gen AI model
application, but with third-party Gen AI Gen AI model with your trained on your data & integrate
limited customization model into your data and integrate it into into your apps for full IP ownership,
and control application(s) your apps control, and customization
3rd Party
3rd Party
3rd Party
3rd Party
In-house
Gen AI Model Gen AI Model Gen AI Model Gen AI Model Gen AI Model
and U se
ChatGPT by O k
penAI is undoubtedly the most well- nown 3rd party Gen AI service nowadays, having
J
reached 100 million monthly active users in anuary, just two months after its launch. Today, the field of
Claude.ai by Anthropic, to more specialized tools such as Git H ub Copilot for drafting code. The following
table shows key aspects of the Pay and Use Generative AI Tools approach. This table will help you compare
the SoftServe approach with other options and better understand if it is right for your business.
L imitations Accuracy, truthfulness, and inherent bias are common issues of Gen AI model
App specific limitations, e.g. ChatGPT cannot access data or current events after
September 2021
The primary limiting factor is that in this business pattern 3rd party Gen AI apps
are not integrated into the corporate ecosystem, which restricts their value to
Proprietary Data Proprietary data might be used in manual prompts or when uploading
Security/Privacy User input may be stored for the purposes of future model training. As for copyright
on the generated content, laws can vary depending on the jurisdiction and the specifics
of the use case, and policies may change over time. For example, under current U.S.
Cost
Typically fixed-fee subscriptions of $10 to $30 per user per month,
Implementation Team This adoption pattern does not require a technical team.
Our recommendation
Evaluate 3rd party Gen AI apps, adopt them for corporate roles where they display value, and monitor their
further development. The field is emerging, and for example, it is anticipated that GPT-5 will be 10 to 20 times
larger than the current GPT-4, which could enable new opportunities for your business.
In the meantime, let’s explore more value that lies in integration with your apps and corporate ecosystem.
Integrate
Your Apps
The good news for businesses is that popular LLM foundational models are available through APIs,
which can be integrated with existing applications or used for building custom software. In the table
below, we will explore key aspects of the “Integrate Your Apps” pattern:
Limitations See limitations in “Pay and Use”. In addition there are technical limitations related to APIs
The size of the context window (the data that the model can "see" or consider) is limited and
varies from service to service. For instance, as of now, GPT-4 supports up to 32K tokens. As a
rough estimate, you can think of 1 token as roughly equivalent to 1 word. This corresponds to
about 100 pages
While this pattern allows for integration with an application's data context, the primary limitation
is that it does not leverage corporate data stores such as wikis, document management, data
lakes, or data warehouses. The next two patterns aim to solve this limitation.
Proprietary Data Proprietary data from corporate apps are passed to a 3rd party
Generative AI service in the form of prompts.
Security/Privacy 3rd party terms can vary but most Cloud vendors do not transfer customer’s
data to other customers and do not use it for improving their models.
Technology Azure OpenAI Service (GPT-4, gpt-3.5-turbo), Google Cloud (PaLM 2), Amazon
Bedrock (Titan, Claude 2, etc.), Nvidia (NeMo). Most services provide REST API and
client libraries for Python, Java, Java Script and other popular languages.
Cost The monthly cost of usage 3rd party Gen AI service heavily depends on the amount of passed
and generated information (tokens), for example a virtual 24/7 agent answering around 10,000
questions per day would cost from $210 to $6,000 monthly (based on GPT-3.5 and GPT-4 Turbo
current pricing). At the same time, there are options for using open-source models that can
noticeably decrease the total cost of infrastructure for high-load applications.
The cost of integration with corporate apps will depend on the number of those apps and their
complexity; thus, it is not included in the calculation.
Our recommendation
Carefully assess your existing applications and identify where Gen AI integration can provide the most value.
Consider the technical limitations, such as the context window size and the cost based on the leveraging selected
model and API usage. Depending on the option, the cost difference can vary by up to 30 times. When the need
arises in the corporate knowledge base, then the next pattern is worth looking into.
Enrich with
Your Data
?
Looking to use existing LLM solutions with your own data No worries, it doesn't necessarily re uire training q
your own LLM. LLMs introduce a cost-effective approach to knowledge utilization called embeddings.
Through AI-powered indexing and search, GenAI systems can access information from corporate systems
such as portals, wikis, data warehouses, data lakes, and other sources of structured or unstructured data.
The integration of LLMs with proprietary data opens the door to leveraging the distinct knowledge amassed
q
by an organization — a resource uni uely their own.
Use case examples The best use cases for this pattern are the ones that rely on corporate data and knowledge,
e.g., Customer Service Automation, Marketing Content Generation, Auto-reporting, Enterprise
Search, Job Specific Assistants/Copilots, and Decision Support systems.
Limitations LLMs enable enhanced content indexing using AI-powered embeddings (indexes).
However, it also introduces the need for an additional data infrastructure for
processing and storing the embeddings
Fitting corporate knowledge into the model's context could be challenging.
However, adding an extra processing logic may help extract concise summaries
from the most relevant documents.
Proprietary Data Additional proprietary data from corporate data sources is passed to an
embedding model for content indexing. Self-hosted open-source embedding
models can be used to secure sensitive data.
Security/Privacy In addition to the embedding model, corporate data is also passed to a 3rd party LLM,
potentially exposing sensitive data to users via an LLM response. Proper guardrails and
response filtering must be implemented to prevent sensitive data exposure.
Technology Azure OpenAI, Amazon Bedrock, Google Cloud Embedding API, Cohere, and Nvidia NeMo are
the most popular 3rd party services for embedding models. HuggingFace Hub provides a wide
range of open-source embedding models. Milvus, FAISS, Chroma, and Weaviate are the most
popular open-source embedding databases, and Pinecone is a good 3rd-party alternative.
Cost “ Y ”
Like in the abovementioned pattern, Integrate our Apps, the monthly cost of usage
depends on the amount of passed and generated information. However, in this case,
indexing data through embeddings also incurs additional costs .
Thus, using open-source models for embedding can drastically decrease the cost when
supporting massive data.
Implementation Team Solution and Data Architects, Software, Data, and ML Engineers, DevOps Engineer.
Our recommendation
Leverage corporate data as a powerful differentiator. Consider utilizing open-source models to mitigate
expenses, especially when dealing with large volumes of data. Implement robust guardrails and response
filtering to prevent potential exposure of sensitive data. In case your system relies on specific domain expertise,
the next approach — Train on your Data — can be a better option.
Train on
Your Data
The most value can be extracted from LLMs when it’s trained on the specific data relevant to the business
and subject domain. This pattern is more expensive to implement but provides more accurate responses
for specific domains or skills. The good news is that the initial model training can be eliminated by using
pre-trained Foundational Models (FMs) and fine-tuning them on the proprietary data. Some third-party
party LLMs allow fine-tuning the model on the customer's data, while others do not. In the latter case, the
model can be fine-tuned on the customer's data using self-hosted open-source FMs.
Use case examples The use cases for specialized skills or specific domain (Legal, Medical, Finance,
etc.) are the best candidates.
Limitations Training and fine-tuning LLMs requires a significant amount of training data and extensive
compute resources (including GPUs/TPUs)
Data quality is a critical factor for the success of this pattern – smaller but higher quality data
brings better results. Additional data curation and governance processes may be required
Many 3rd-party LLMs are limited in training data customization (e.g., GPT-3.5 and GPT-4 do
not support fine-tuning on the customer's data, only older versions)
LLMs trained on a specific domain may not be able to generalize to other domains.
Proprietary Data In addition to user and corporate search data, fine-tuning a 3rd party LLM requires passing a
significant amount of data relevant to the business and subject domain for knowledge extraction.
Security/Privacy 3rd party terms may vary but most Cloud vendors do not share customer’s
data and fined-tuned models between tenants.
Technology Google Vertex AI, Amazon SageMaker, Azure Machine Learning, HuggingFace Hub,
and Nvidia NeMo are the most popular services for training and fine-tuning LLMs.
HuggingFace Hub provides a wide range of open-source Foundational Models.
Cost The cost of fine-tuning varies heavily depending on the approach: managed service vs. open
source. For example, fine-tuning foundational models with biomedical data (~50B tokens)
using the OpenAI Davinci model costs $1,500,000, compared to $43,550 when using the
open-source MPT-30B. For comparison with fine-tuning, here are some examples of training
foundational models from scratch:
Implementation Team Solution and Data Architects, Data Scientist, Software, Data, and ML Engineers, DevOps Engineer
Our recommendation
The “Train on Your Data”approach is more complex and expensive but offers the greatest benefits. Leverage pre-trained
foundational models and fine-tune them on proprietary data to reduce initial model training costs. Pay attention to data
quality and consider additional curation and governance processes, as smaller but higher quality data may yield better
results, and be mindful of the significant costs and resources required for training from scratch.
Get Started on
your Generative AI
Journey
Let SoftServe Accelerate Your
Generative AI Journey with 3 Offerings
Don't be left behind in the race to harness the disruptive innovation of Generative AI. SoftServe's adoption
patterns are designed to help you navigate the complexities of this emerging technology. We will help you
select an appropriate adoption pattern or its combinations for building tailored solutions to meet your
enterprise's unique needs. Expedite your journey towards a new way of productivity and creativity with
SoftServe Generative AI offerings, positioning your business as a leader in the industry:
with POC
Generative AI Solution
Development
Generative AI in Product
and Engineering Teams
It's urgent and the time to start is now. But you know that Generative AI is not a deployment sprint. With
SoftServe's experts to address your unique enterprise needs and guide you through possible adoption
options, you will move past the hype and harness its benefits for your enterprise.
Learn more about SoftServe’s Generative AI Lab, POVs, offerings, and partners.
UK Sweden
Germany Poland
Canada Bulgaria Ukraine
Romania
USA
UAE
Mexico
Singapore
Colombia
Chile
Microsoft
Innovate with purpose, rationalize costs, and drive efficiencies with Microsoft Azure's open and
flexible cloud computing platform. Leverage Azure OpenAI and Machine Learning services to deliver
next-generation AI solutions. As a Gold Microsoft Partner, SoftServe enables your business to build
and deploy on your terms — both today and in the future. Learn more
250+ 19+
Azure Certified
Years as a
N vidia
As an NVIDIA Service Delivery Partner, SoftServe harnesses NVIDIA's cutting-edge technologies, like
GPU-accelerated compute infrastructure, to deliver robust AI solutions. Leveraging NVIDIA's NeMo
Service, SoftServe streamlines the development of Generative AI products, driving rapid digital
transformation. Learn more
deep technology expertise in AI/ML
Iurii Milovanov,
Serge Haziyev,
Alex Chubay,