Why Upstage Builds Small Language Models

Share via:


LAS VEGAS — Upstage is a South Korean enterprise AI company that builds small language models (SLMs)to help companies solve document processing problems. It originated as a company using optical character recognition (OCR) to scan documents for large corporations in South Korea.

When ChatGPT emerged, customers started asking Upstage about large language models (LLMs). Upstage had provided 95% accuracy using its OCR capability, but customers wanted 100% accuracy. So, the Upstate team began looking at models that would fit the requirements for getting better accuracy. The LLMs serve a general purpose, but the smaller models were more applicable to the narrow focus that document processing requires.

There’s not much attention paid to SLMs, but their capabilities have uses that include providing company-specific or even country-specific LLMs.

“Customers wanted a language model that was fit for their own use,” said Lucy Park, co-founder and chief product officer, in an interview at AWS re:Invent. “So that’s one of the reasons we started out to build small language models. And so here we are working on document processing engines and large language models.”‘

Model Merging to Create SLMs

Upstage, an AWS Generative AI Accelerator participant uses open source models, allowing running on a single GPU. Its flagship model, Solar, compares to other small models that also run on a single GPU, including Llama 3.81 B, Mistral Small Instruct 2409, and Hugging Face’s ExaOne3.0 7.8B Instruct.

Park said Upstage merges two copies of a small LLM into a large LLM. For instance, it would integrate a 7 billion parameter model into a 10 billion parameter mode. “If we have a 14 billion model, we explode that into a 22 billion model,” she said. “So that’s what we have been doing recently.”

Model merging, a technique for combining LLMs has gained acceptance in the AI community.  Implementation includes such practices as weight averaging, a method that merges the parameters of multiple separate models with different capabilities. Model merging allows data scientists “to build a universal model without needing access to the original training data or expensive computation,” according to a paper published in August by researchers from Nanyang Technological University, Northeastern University and Sun Yat-sen University.

Park said Upstage has found increases in its benchmarks using a combined model approach. According to the Upstage site, Solar Pro is a small language model that shows a 64% improvement in Eastern Asia language mastery compared to Solar Pro preview.

The improvements in SLMs for languages reflect their growing popularity. SLMs train smaller data sets, making them flexible for domain-centered approaches like Upstage’s.

Park said the large language models focus on general intelligence. The small language models also provide a narrower focus.

For example, Upside built a specific model for the Thai language. With Thai, it’s similar to GPT 4, the OpenAI model.

SLMs also cost a lot less to develop. Hypothetically, Park said, imagine an SLM that costs $10 to build. An LLM that is 10 times bigger may cost $100.

Customers will pursue three options to deploy the models, she said. If they are deploying on-premise models, they can use the Upstage console, which provides APIs through the AWS marketplace. For example, the Solar Pro model is now available on the Amazon Bedrock Marketplace.


Group Created with Sketch.

ath d=”M24.002,29.619 L29.77,29.619 L29.77,15.808 C29.77,15.038 29.622,11.265 29.59,10.414 L29.77,10.414 C31.424,14.019 31.473,14.147 32.168,15.322 L39.65,29.618 L44.845,29.618 L44.845,0 L39.075,0 L39.075,11.064 C39.075,12.197 39.075,12.44 39.182,14.472 L39.325,17.468 L39.151,17.468 C39.034,17.267 38.596,16.173 38.467,15.929 C38.164,15.323 37.725,14.512 37.373,13.905 L30.031,0 L24,0 L24,29.619 L24.002,29.619 Z” id=”Path-Copy” fill=”#FF3287″/>

ath d=”M56.948,0 C50.745,0 47.606,3.43 47.606,8.296 C47.606,14.114 51.036,15.404 55.518,17.132 C60.438,18.853 61.782,19.332 61.782,21.539 C61.782,24.225 58.969,24.867 57.401,24.867 C54.579,24.867 52.493,23.342 51.536,20.858 L47,24.185 C49.43,28.937 52.145,30.185 57.713,30.185 C59.364,30.185 62.059,29.74 63.727,28.694 C67.779,26.156 67.779,22.22 67.779,20.898 C67.779,18.129 66.531,16.207 66.178,15.726 C65.049,14.121 63.032,12.918 61.25,12.278 L57.084,10.914 C55.073,10.267 52.928,10.105 52.928,8.019 C52.928,7.707 53.008,5.528 56.288,5.319 L61.465,5.319 L61.465,0 C61.465,0 57.342,0 56.948,0 Z” id=”Path-Copy-2″ fill=”#00AFF4″/>

olygon id=”Path” fill=”#00AFF4″ points=”5.32907052e-15 1.77635684e-15 5.32907052e-15 5.319 7.572 5.319 7.572 29.564 14.132 29.564 14.132 5.319 21.544 5.319 21.544 1.77635684e-15″/>





Source link

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Team SNFYI
Hi! This is Admin.

Popular

More Like this

Why Upstage Builds Small Language Models


LAS VEGAS — Upstage is a South Korean enterprise AI company that builds small language models (SLMs)to help companies solve document processing problems. It originated as a company using optical character recognition (OCR) to scan documents for large corporations in South Korea.

When ChatGPT emerged, customers started asking Upstage about large language models (LLMs). Upstage had provided 95% accuracy using its OCR capability, but customers wanted 100% accuracy. So, the Upstate team began looking at models that would fit the requirements for getting better accuracy. The LLMs serve a general purpose, but the smaller models were more applicable to the narrow focus that document processing requires.

There’s not much attention paid to SLMs, but their capabilities have uses that include providing company-specific or even country-specific LLMs.

“Customers wanted a language model that was fit for their own use,” said Lucy Park, co-founder and chief product officer, in an interview at AWS re:Invent. “So that’s one of the reasons we started out to build small language models. And so here we are working on document processing engines and large language models.”‘

Model Merging to Create SLMs

Upstage, an AWS Generative AI Accelerator participant uses open source models, allowing running on a single GPU. Its flagship model, Solar, compares to other small models that also run on a single GPU, including Llama 3.81 B, Mistral Small Instruct 2409, and Hugging Face’s ExaOne3.0 7.8B Instruct.

Park said Upstage merges two copies of a small LLM into a large LLM. For instance, it would integrate a 7 billion parameter model into a 10 billion parameter mode. “If we have a 14 billion model, we explode that into a 22 billion model,” she said. “So that’s what we have been doing recently.”

Model merging, a technique for combining LLMs has gained acceptance in the AI community.  Implementation includes such practices as weight averaging, a method that merges the parameters of multiple separate models with different capabilities. Model merging allows data scientists “to build a universal model without needing access to the original training data or expensive computation,” according to a paper published in August by researchers from Nanyang Technological University, Northeastern University and Sun Yat-sen University.

Park said Upstage has found increases in its benchmarks using a combined model approach. According to the Upstage site, Solar Pro is a small language model that shows a 64% improvement in Eastern Asia language mastery compared to Solar Pro preview.

The improvements in SLMs for languages reflect their growing popularity. SLMs train smaller data sets, making them flexible for domain-centered approaches like Upstage’s.

Park said the large language models focus on general intelligence. The small language models also provide a narrower focus.

For example, Upside built a specific model for the Thai language. With Thai, it’s similar to GPT 4, the OpenAI model.

SLMs also cost a lot less to develop. Hypothetically, Park said, imagine an SLM that costs $10 to build. An LLM that is 10 times bigger may cost $100.

Customers will pursue three options to deploy the models, she said. If they are deploying on-premise models, they can use the Upstage console, which provides APIs through the AWS marketplace. For example, the Solar Pro model is now available on the Amazon Bedrock Marketplace.


Group Created with Sketch.

ath d=”M24.002,29.619 L29.77,29.619 L29.77,15.808 C29.77,15.038 29.622,11.265 29.59,10.414 L29.77,10.414 C31.424,14.019 31.473,14.147 32.168,15.322 L39.65,29.618 L44.845,29.618 L44.845,0 L39.075,0 L39.075,11.064 C39.075,12.197 39.075,12.44 39.182,14.472 L39.325,17.468 L39.151,17.468 C39.034,17.267 38.596,16.173 38.467,15.929 C38.164,15.323 37.725,14.512 37.373,13.905 L30.031,0 L24,0 L24,29.619 L24.002,29.619 Z” id=”Path-Copy” fill=”#FF3287″/>

ath d=”M56.948,0 C50.745,0 47.606,3.43 47.606,8.296 C47.606,14.114 51.036,15.404 55.518,17.132 C60.438,18.853 61.782,19.332 61.782,21.539 C61.782,24.225 58.969,24.867 57.401,24.867 C54.579,24.867 52.493,23.342 51.536,20.858 L47,24.185 C49.43,28.937 52.145,30.185 57.713,30.185 C59.364,30.185 62.059,29.74 63.727,28.694 C67.779,26.156 67.779,22.22 67.779,20.898 C67.779,18.129 66.531,16.207 66.178,15.726 C65.049,14.121 63.032,12.918 61.25,12.278 L57.084,10.914 C55.073,10.267 52.928,10.105 52.928,8.019 C52.928,7.707 53.008,5.528 56.288,5.319 L61.465,5.319 L61.465,0 C61.465,0 57.342,0 56.948,0 Z” id=”Path-Copy-2″ fill=”#00AFF4″/>

olygon id=”Path” fill=”#00AFF4″ points=”5.32907052e-15 1.77635684e-15 5.32907052e-15 5.319 7.572 5.319 7.572 29.564 14.132 29.564 14.132 5.319 21.544 5.319 21.544 1.77635684e-15″/>





Source link

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Website Upgradation is going on for any glitch kindly connect at office@startupnews.fyi

Team SNFYI
Hi! This is Admin.

More like this

Whack-a-mole: US academic fights to purge his AI deepfakes

As deepfake videos of John Mearsheimer multiplied across...

WhatsApp beta for Android 2.26.5.1: what’s new?

WhatsApp has released a new Android update through...

WordPress Publishes AI Guidelines To Combat AI Slop

WordPress published guidelines for using AI for coding...

Popular

iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista melhor iptv portugal lista best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv best iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv portugal iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv iptv