Skip to content
Tiatra, LLCTiatra, LLC
Tiatra, LLC
Information Technology Solutions for Washington, DC Government Agencies
  • Home
  • About Us
  • Services
    • IT Engineering and Support
    • Software Development
    • Information Assurance and Testing
    • Project and Program Management
  • Clients & Partners
  • Careers
  • News
  • Contact
 
  • Home
  • About Us
  • Services
    • IT Engineering and Support
    • Software Development
    • Information Assurance and Testing
    • Project and Program Management
  • Clients & Partners
  • Careers
  • News
  • Contact

What LinkedIn learned leveraging LLMs for its billion users

With more than 1 billion users globally, LinkedIn is continuously bumping against the limits of what is technically feasible in the enterprise today. Few companies operate at the scale that LinkedIn does or have access to similar troves of data.

For the business- and employment-focused social media platform, connecting qualified candidates with potential employers to help fill job openings is core business. So too is ensuring that post feeds across the platform are relevant to the members who consume them. And at LinkedIn scale, those matchmaking processes have always relied on technology.

During the summer of 2023, at the height of the first wave of interest in generative AI, LinkedIn began to wonder whether matching candidates with employers and making feeds more useful would be better served with the help of large language models (LLMs).

So the social media giant launched a generative AI journey and is now reporting the results of its experience leveraging Microsoft’s Azure OpenAI Service. CIOs in every vertical can take a tip or two from the lessons LinkedIn learned along the way.

Fits and starts

As most CIOs have experienced, embracing emerging technologies comes with its share of experimentation and setbacks. For LinkedIn, this was no different, as its road to LLM insights was anything but smooth, said LinkedIn’s Juan Bottaro, a principal software engineer and tech lead.

The initial deliverables “felt lacking,” Bottaro said. “Not enough dots were being connected.”

Those first waves of hype around generative AI didn’t help.

“LLM was the new thing and it felt like it could solve everything,” Bottaro said. “We didn’t start with a very clear idea of what an LLM could do.”

For example, an early version of the revised job-matching effort was rather, for the lack of a better word, rude. Or at least overly blunt.

“Clicking ‘Assess my fit for this job’ and getting ‘You are a terrible fit’ isn’t very useful,” Bottaro said. “We wanted [responses] to be factual but also empathetic. Some members may be contemplating a career change into fields where they currently do not have a strong fit and need help understanding what are the gaps and the next steps.”

So, among LinkedIn’s first lessons was the importance of tuning LLMs to audience expectations — and helping the LLMs understand how to be perhaps not human, but at least humane with its responses.

A question of speed

Even though LinkedIn has more than a billion members, most of the job search features that would rely on LinkedIn’s LLM work were to be targeted initially to Premium members, a far smaller group. (LinkedIn declined to comment on how many Premium members it has.) 

When operating at that scale, speed can be essential, especially for something as nuanced as matching candidates with relevant job openings. Here, it was believed an LLM would help, as an oft-touted benefit of LLMs is their speed, enabling them to complete complex steps rapidly. That was notthe case with LinkedIn’s deployment, Bottaro said.

“I wouldn’t characterize LLMs as fast. I wouldn’t characterize speed as an advantage,” he said.

But speed can be defined in multiple ways. Even though operationally the LLM may not have been as fast as hoped, Bottaro said the acceleration of the overall deployment process was astounding. “The superpower of this new technology is that you can create prototypes very fast, somewhere between two and three months. That was not possible before this technology,” he said. 

Asked how long the various facets of the project would have taken without LLM, Bottaro said some couldn’t have been accomplished at all while other elements “maybe it would have taken several years.”

As an example, Bottaro referenced the part of the system designed to understand intent. Without an LLM, that might have taken two to three months, he said, but the LLM mastered it “in less than a week.”

Cost considerations

One aspect that Bottaro dubbed “a hurdle” was the cost. Again, cost can mean different things in different phases of a project, as LinkedIn’s experience shows.

“The amount that we spent to be able to develop was negligible,” Bottaro said. But when it came to delivering data to LinkedIn’s customers, costs exploded.

“Even if it was just going to a couple million members,” said Bottaro, potentially hinting at the number of Premium users, the pricing soared. That is because LLM pricing — at least the licensing deal that LinkedIn struck with Microsoft, its LLM provider and parent company — was based on usage, specifically input and output tokens.

One AI vendor CEO, Tarun Thummala, explains in a LinkedIn post unrelated to this project that LLM input and output tokens are roughly equivalent to 0.75 of a word. LLM vendors typically sell tokens by the thousands or millions. Azure OpenAI, which LinkedIn uses, charges $30 for every 1 million 8K GPT-4 input tokens and $60 for every 1 million 8K GPT-4 output tokens out of its East US region, for example.

Evaluation challenges

Another functionality goal LinkedIn had for its project was automatic evaluation. LLMs are notoriously challenging to assess in terms of accuracy, relevancy, safety, and other concerns. Leading organizations, and LLM makers, have been attempting to automate some of this work, but according to LinkedIn, such capabilities are “still a work in progress.”

Without automated evaluation, LinkedIn reports that “engineers are left eye-balling results and testing on a limited set of examples and having a more than a 1+ day delay to know metrics.”

The company is building model-based evaluators to help estimate key LLM metrics, such as overall quality score, hallucination rate, coherence, and responsible AI violations. Doing so will enable faster experimentation, the company’s engineers say, and though LinkedIn’s engineers have had some success with hallucination detection, they haven’t been able to finish work in that area yet.

Data quality

Part of the struggle LinkedIn experienced with its job match effort boils down to a data quality issue from both sides: employers and potential employees.

LLMs can work only with what it is given and sometimes job posts are not precise or comprehensive in what skills an employer seeks. On the flip side, some job applicants post poorly phrased histories that do not effectively reflect their wealth of experience in problem-solving, for example. 

Here, Bottaro sees potential for LLMs to help employers and potential employees alike. By improving what employers and LinkedIn users write, both sides could benefit, as the company’s job-matching LLM could work far more effectively when its data inputs are higher quality.

User experience

When dealing with such a large member base, accuracy and relevancy metrics can “give a false sense of comfort,” Bottaro said. For example, if the LLM “gets it right 90% of the time, that means that one out of ten would have a horrible experience,” he said.

Making this deployment more difficult is the extreme level of nuance and judgment involved in delivering useful, helpful, and accurate answers.

“How do you agree what is good and bad? We spent a lot of time with linguists to make guidelines about how to deliver a thorough representation. We did a lot of user studies, too,” Bottaro said. “How do you train the person to write the right response? How do you define the task, specify how the response should look? The product may be trying to be constructive or helpful. It is not trying to assume too much because thatis where hallucinations start. The consistency of the responses is something we are super proud of.”

Operating in real-time

LinkedIn’s sheer scale presents another challenge for job matches. With a billion members, a job ad can have hundreds — if not thousands — of responses within a few minutes of posting. Many job applicants won’t bother to apply if they see that hundreds have already applied. That puts the onus on the LLM to very quickly find members that are matches before less qualified applicants submit their material. After that, it’s still a matter of that member seeing the notification and reacting to it quickly enough.

On the employer side, the struggle is finding the applicants that are the best fit — not necessarily those who respond the quickest. The reluctance of some companies to post salary ranges further complicates both efforts as the best qualified applicants may not be interested in a role depending on its compensation. That is a problem that an LLM can’t fix.

APIs and RAG

LinkedIn’s massive trove of data includes a lot of unique information about individuals, employers, skills, and coursework, which its LLMs have not been trained on. As a result, its LLMs are unable to use these assets, as they are currently stored and delivered, for any of their reasoning or response-generating activities, according to LinkedIn’s engineers.

Here, retrieval augmented generation (RAG) is a typical work-around. By setting up a pipeline of internal APIs, enterprises can ‘augment’ LLM prompts with additional context to better guide — and guardrail — the LLM’s response. Much of LinkedIn’s data is exposed via RPC APIs, something the company’s engineers say is “convenient for humans to invoke programmatically,” but “is not very LLM friendly.”

To solve this issue, LinkedIn’s engineers “wrapped skills” around its APIs to give them an “LLM friendly description of what the API does and when to use it,” as well as configuration details, input and output schema, and all the necessary logic to map the LLM version of each API to its underlying (actual) RPC one, LinkedIn said.

“Skills like this enable the LLM to do various things relevant to our product like view profiles,

Search articles/people/jobs/companies and even query internal analytics systems,” the company’s engineers wrote in a statement. “The same technique is also used for calling non-LinkedIn APIs like Bing search and news.”

Generative AI, Software Development, Technology Industry


Read More from This Article: What LinkedIn learned leveraging LLMs for its billion users
Source: News

Category: NewsApril 25, 2024
Tags: art

Post navigation

PreviousPrevious post:Oracle adds AI capabilities to its Fusion Cloud CXNextNext post:IBM doubles down on hybrid cloud with $6.4B HashiCorp acquisition

Related posts

휴먼컨설팅그룹, HR 솔루션 ‘휴넬’ 업그레이드 발표
May 9, 2025
Epicor expands AI offerings, launches new green initiative
May 9, 2025
MS도 합류··· 구글의 A2A 프로토콜, AI 에이전트 분야의 공용어 될까?
May 9, 2025
오픈AI, 아시아 4국에 데이터 레지던시 도입··· 한국 기업 데이터는 한국 서버에 저장
May 9, 2025
SAS supercharges Viya platform with AI agents, copilots, and synthetic data tools
May 8, 2025
IBM aims to set industry standard for enterprise AI with ITBench SaaS launch
May 8, 2025
Recent Posts
  • 휴먼컨설팅그룹, HR 솔루션 ‘휴넬’ 업그레이드 발표
  • Epicor expands AI offerings, launches new green initiative
  • MS도 합류··· 구글의 A2A 프로토콜, AI 에이전트 분야의 공용어 될까?
  • 오픈AI, 아시아 4국에 데이터 레지던시 도입··· 한국 기업 데이터는 한국 서버에 저장
  • SAS supercharges Viya platform with AI agents, copilots, and synthetic data tools
Recent Comments
    Archives
    • May 2025
    • April 2025
    • March 2025
    • February 2025
    • January 2025
    • December 2024
    • November 2024
    • October 2024
    • September 2024
    • August 2024
    • July 2024
    • June 2024
    • May 2024
    • April 2024
    • March 2024
    • February 2024
    • January 2024
    • December 2023
    • November 2023
    • October 2023
    • September 2023
    • August 2023
    • July 2023
    • June 2023
    • May 2023
    • April 2023
    • March 2023
    • February 2023
    • January 2023
    • December 2022
    • November 2022
    • October 2022
    • September 2022
    • August 2022
    • July 2022
    • June 2022
    • May 2022
    • April 2022
    • March 2022
    • February 2022
    • January 2022
    • December 2021
    • November 2021
    • October 2021
    • September 2021
    • August 2021
    • July 2021
    • June 2021
    • May 2021
    • April 2021
    • March 2021
    • February 2021
    • January 2021
    • December 2020
    • November 2020
    • October 2020
    • September 2020
    • August 2020
    • July 2020
    • June 2020
    • May 2020
    • April 2020
    • January 2020
    • December 2019
    • November 2019
    • October 2019
    • September 2019
    • August 2019
    • July 2019
    • June 2019
    • May 2019
    • April 2019
    • March 2019
    • February 2019
    • January 2019
    • December 2018
    • November 2018
    • October 2018
    • September 2018
    • August 2018
    • July 2018
    • June 2018
    • May 2018
    • April 2018
    • March 2018
    • February 2018
    • January 2018
    • December 2017
    • November 2017
    • October 2017
    • September 2017
    • August 2017
    • July 2017
    • June 2017
    • May 2017
    • April 2017
    • March 2017
    • February 2017
    • January 2017
    Categories
    • News
    Meta
    • Log in
    • Entries feed
    • Comments feed
    • WordPress.org
    Tiatra LLC.

    Tiatra, LLC, based in the Washington, DC metropolitan area, proudly serves federal government agencies, organizations that work with the government and other commercial businesses and organizations. Tiatra specializes in a broad range of information technology (IT) development and management services incorporating solid engineering, attention to client needs, and meeting or exceeding any security parameters required. Our small yet innovative company is structured with a full complement of the necessary technical experts, working with hands-on management, to provide a high level of service and competitive pricing for your systems and engineering requirements.

    Find us on:

    FacebookTwitterLinkedin

    Submitclear

    Tiatra, LLC
    Copyright 2016. All rights reserved.