{"id":236938,"date":"2025-02-28T13:30:04","date_gmt":"2025-02-28T05:30:04","guid":{"rendered":"https:\/\/www.grab.com\/sg\/?post_type=editorial&#038;p=236938"},"modified":"2025-12-01T15:22:56","modified_gmt":"2025-12-01T07:22:56","slug":"how-large-language-models-help-us-make-more-precise-content-moderation-decisions","status":"publish","type":"editorial","link":"https:\/\/www.grab.com\/sg\/inside-grab\/stories\/how-large-language-models-help-us-make-more-precise-content-moderation-decisions\/","title":{"rendered":"How Large Language Models help us make more precise content moderation decisions"},"content":{"rendered":"\t\t<div data-elementor-type=\"wp-post\" data-elementor-id=\"236938\" class=\"elementor elementor-236938\" data-elementor-post-type=\"editorial\">\n\t\t\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-a418b45 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"a418b45\" data-element_type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-ead919e\" data-id=\"ead919e\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap\">\n\t\t\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-2d95621 elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"2d95621\" data-element_type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-6d649a9 gr21-boxed-content editorial-gr21-boxed-content\" data-id=\"6d649a9\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap elementor-element-populated\">\n\t\t\t\t\t\t<div class=\"elementor-element elementor-element-782c453 elementor-widget elementor-widget-text-editor\" data-id=\"782c453\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<p>At Grab, we want to provide a positive and safe experience for everyone using our app. Therefore, we need a solid content moderation system that helps us screen user-generated content\u2014such as merchant catalogues and user reviews\u2014to flag incidents of offensive language, inappropriate images, and so on.<\/p><p>Thankfully, the vast majority of interactions and content uploaded to our platform is harmless. And with good content moderation, we can catch the bad apples early on.<\/p><p>We\u2019ve always done this with a blend of AI-supported automated content filters and human moderators weighing in on the more complex decisions. More recently, however, the rise of Large Language Models (LLMs) made our automated systems more precise, which means we\u2019re getting more effective at quick, accurate content moderation at a mass-scale, while reducing the strain on our content moderation team. Here\u2019s how<\/p>\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-1c1ed31 elementor-widget elementor-widget-image\" data-id=\"1c1ed31\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"700\" height=\"340\" src=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-700x340.png\" class=\"attachment-large size-large wp-image-236972\" alt=\"\" srcset=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-700x340.png 700w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-250x121.png 250w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-768x373.png 768w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-1536x745.png 1536w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-18x9.png 18w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table-120x58.png 120w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28130401\/Content-moderation-at-Grab-table.png 1562w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-5ea5621 elementor-widget elementor-widget-text-editor\" data-id=\"5ea5621\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<h5>Content moderation is getting better thanks to LLMs<\/h5><p>Grab\u2019s content moderation works in a two-tier system. In the first layer, we employ small, specialised AI models that can screen large amounts of content quickly and efficiently.<\/p><p>An example of a Tier 1 model is keyword filtration. Based on a pre-defined list of problematic words, the system flags content that contains any of these words. Similarly, an AI model can screen images for potential violations after it has been pre-trained.<\/p><p>Such task-specific AIs manage to \u201cpass\u201d or \u201cfail\u201d large amounts of data with high accuracy. Our Tier 1 models currently flag less than 5% of content as potential violations. These then get passed on to \u201clarge\u201d Tier 2 models for another assessment<\/p>\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-db567ff elementor-widget elementor-widget-image\" data-id=\"db567ff\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"512\" height=\"253\" src=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28131217\/AI-content-moderation-tiers-flow.png\" class=\"attachment-large size-large wp-image-236979\" alt=\"\" srcset=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28131217\/AI-content-moderation-tiers-flow.png 512w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28131217\/AI-content-moderation-tiers-flow-250x124.png 250w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28131217\/AI-content-moderation-tiers-flow-18x9.png 18w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28131217\/AI-content-moderation-tiers-flow-120x59.png 120w\" sizes=\"(max-width: 512px) 100vw, 512px\" \/>\t\t\t\t\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-52ccf63 elementor-widget elementor-widget-text-editor\" data-id=\"52ccf63\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<h5>Context-aware decision-making with LLMs<\/h5><p>In Tier 2, LLMs can make more complex, context-aware decisions. We can simply prompt the LLM with our violation policies on specific topics, let\u2019s say the sale of tobacco on Grab, or our sexual harassment policy. This applies even to complex scenarios, for example to differentiate traditional tobacco and e-cigarettes. The prompts for detecting all tobacco and the prompt for detecting e-cigarette are shown in the table below. This gives a high level of flexibility to handle complex scenarios defined in policies.<\/p>\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-e927ffb elementor-widget elementor-widget-image\" data-id=\"e927ffb\" data-element_type=\"widget\" data-widget_type=\"image.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t\t\t\t<figure class=\"wp-caption\">\n\t\t\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"700\" height=\"330\" src=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-700x330.png\" class=\"attachment-large size-large wp-image-236987\" alt=\"\" srcset=\"https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-700x330.png 700w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-250x118.png 250w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-768x362.png 768w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-18x8.png 18w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example-120x57.png 120w, https:\/\/assets.grab.com\/wp-content\/uploads\/sites\/4\/2025\/02\/28132140\/Grab-content-moderation-policy-prompt-example.png 784w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/>\t\t\t\t\t\t\t\t\t\t\t<figcaption class=\"widget-image-caption wp-caption-text\">Content moderation prompt examples<\/figcaption>\n\t\t\t\t\t\t\t\t\t\t<\/figure>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t<div class=\"elementor-element elementor-element-c6d4056 elementor-widget elementor-widget-text-editor\" data-id=\"c6d4056\" data-element_type=\"widget\" data-widget_type=\"text-editor.default\">\n\t\t\t\t<div class=\"elementor-widget-container\">\n\t\t\t\t\t\t\t<p><span style=\"font-weight: 400;\">The LLM can then assess how likely it is that a piece of content is in violation of one of our policies. A low score means the LLM is confident the content is safe. Meanwhile, a high score indicates the content should be filtered. A medium score means the LLM isn\u2019t quite sure\u2014it could go either way. That\u2019s where human moderators step in for the final say.<\/span><\/p><p><span style=\"font-weight: 400;\">This second layer of automated moderation is where we\u2019ve seen the most change. We started implementing LLMs in Tier 2 moderation in 2023 and are gradually expanding their use to additional cases through Q3 2024. Thus we are able to reduce 90% of human efforts in the moderation process and reduce SLA from days to minutes.<\/span><\/p><h5>Next steps: finetuning our models<\/h5><p><span style=\"font-weight: 400;\">We achieved great results leveraging LLMs, but we\u2019re still working on improvements, specifically in two areas:<\/span><\/p><ul><li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Latency: The LLM latency with image input is still quite high. We therefore need to improve latency to achieve use cases in which users get near-instant feedback about their uploaded content<\/span><\/li><li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">Accuracy: LLMs don\u2019t always have good knowledge on the topics we want to detect. Also, understanding local languages can be challenging. We therefore need to enhance the model\u2019s understanding of those areas<\/span><\/li><\/ul><p><span style=\"font-weight: 400;\">Leveraging our in-house data, our team is constantly finetuning our content moderation LLMs to address the challenges above, with promising initial results.<\/span><\/p>\t\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<section class=\"elementor-section elementor-top-section elementor-element elementor-element-202737c elementor-section-boxed elementor-section-height-default elementor-section-height-default\" data-id=\"202737c\" data-element_type=\"section\">\n\t\t\t\t\t\t<div class=\"elementor-container elementor-column-gap-default\">\n\t\t\t\t\t<div class=\"elementor-column elementor-col-100 elementor-top-column elementor-element elementor-element-e44bac9\" data-id=\"e44bac9\" data-element_type=\"column\">\n\t\t\t<div class=\"elementor-widget-wrap\">\n\t\t\t\t\t\t\t<\/div>\n\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t<\/section>\n\t\t\t\t<\/div>\n\t\t","protected":false},"parent":180237,"menu_order":0,"template":"grab21-default","acf":[],"_links":{"self":[{"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/editorial\/236938"}],"collection":[{"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/editorial"}],"about":[{"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/types\/editorial"}],"version-history":[{"count":32,"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/editorial\/236938\/revisions"}],"predecessor-version":[{"id":254927,"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/editorial\/236938\/revisions\/254927"}],"up":[{"embeddable":true,"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/editorial\/180237"}],"wp:attachment":[{"href":"https:\/\/www.grab.com\/sg\/wp-json\/wp\/v2\/media?parent=236938"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}