Characterization of SLM Conversational Systems Models, Overview

Mir, Yuniesky Orlando Vasconcelo; Pupo, Iliana Pérez; Herrera, Raykenler Yzquierdo; Pérez, Pedro Yobanis Piñero; Acuña, Luis Alvarado; Pérez, Rafael Bello

Abstract

The development of Generative Artificial Intelligence is revolutionizing human–machine interaction models. In this context, large language models (LLMs) have emerged as tools capable of learning complex patterns. However, many of these technologies are highly resource intensive. An alternative to these complex models is the advent of Small Language Models (SLMs). These smaller models process fewer parameters but achieve acceptable performance in their responses, striking a balance between cost and quality. This study characterizes different SLMs to facilitate decision-making in their implementation. In the methods section, a systematic review is conducted, serving as a guide for researchers and professionals seeking to select the most suitable SLM for their specific needs. An analysis of the efficiency of these models contributes to the application of Artificial Intelligence techniques from a sustainability perspective. The results section presents a comparison of various SLMs available on the Ollama platform. The models compared include Qwen2.5, Phi3.5, Mistral-small, Llama3.1, and Gemma2. A comparative analysis evaluates these models based on their efficiency and effectiveness in terms of computational resources and the human effort required to develop task-specific conversational systems. The study demonstrates the feasibility of using these smaller models in various decision-making environments.

Más información

Título según SCOPUS: ID SCOPUS_ID:105001389622 Not found in local SCOPUS DB
Título de la Revista: Studies in Computational Intelligence
Volumen: 1195
Fecha de publicación: 2025
Página de inicio: 3
Página final: 31
DOI:

10.1007/978-3-031-83643-5_1

Notas: SCOPUS